[ 460.712752] env[62368]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62368) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 460.713099] env[62368]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62368) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 460.713228] env[62368]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62368) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 460.713505] env[62368]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 460.807619] env[62368]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62368) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 460.818174] env[62368]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62368) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 461.417728] env[62368]: INFO nova.virt.driver [None req-26df8aef-546c-43e7-8cfb-26ab05b4f641 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 461.491411] env[62368]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 461.491566] env[62368]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 461.491664] env[62368]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62368) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 464.598308] env[62368]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-ebf240f2-d6f6-4654-b783-26a88aa186f4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.614284] env[62368]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62368) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 464.614480] env[62368]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-b0845737-9d4c-450f-9ab6-81eb7d67b90f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.646325] env[62368]: INFO oslo_vmware.api [-] Successfully established new session; session ID is dfb58. [ 464.646501] env[62368]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.155s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 464.647010] env[62368]: INFO nova.virt.vmwareapi.driver [None req-26df8aef-546c-43e7-8cfb-26ab05b4f641 None None] VMware vCenter version: 7.0.3 [ 464.650337] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b63ed8d-6d45-46cf-9968-da983011b6dd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.667320] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ba4b34-4f6b-4774-9dd0-3d2f403fc67b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.672986] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4e334b-81df-4145-8328-53f04b2ce07d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.679322] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65da5e0-390b-43cd-b59f-b81c27858114 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.691899] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cac4415-9971-46fb-94e1-60298739a356 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.697478] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9d8df1-9d76-4734-a986-1d105293de41 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.726631] env[62368]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-85a6fde9-0dd9-45d7-bab8-81b61f6182f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.731314] env[62368]: DEBUG nova.virt.vmwareapi.driver [None req-26df8aef-546c-43e7-8cfb-26ab05b4f641 None None] Extension org.openstack.compute already exists. {{(pid=62368) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 464.734243] env[62368]: INFO nova.compute.provider_config [None req-26df8aef-546c-43e7-8cfb-26ab05b4f641 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 465.238074] env[62368]: DEBUG nova.context [None req-26df8aef-546c-43e7-8cfb-26ab05b4f641 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),d117b194-c7ff-4d07-bcc2-1fe5b28965c2(cell1) {{(pid=62368) load_cells /opt/stack/nova/nova/context.py:464}} [ 465.239765] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 465.239986] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 465.240719] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 465.241187] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Acquiring lock "d117b194-c7ff-4d07-bcc2-1fe5b28965c2" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 465.241349] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Lock "d117b194-c7ff-4d07-bcc2-1fe5b28965c2" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 465.242394] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Lock "d117b194-c7ff-4d07-bcc2-1fe5b28965c2" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 465.262235] env[62368]: INFO dbcounter [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Registered counter for database nova_cell0 [ 465.270257] env[62368]: INFO dbcounter [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Registered counter for database nova_cell1 [ 465.273539] env[62368]: DEBUG oslo_db.sqlalchemy.engines [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62368) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 465.273896] env[62368]: DEBUG oslo_db.sqlalchemy.engines [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62368) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 465.278768] env[62368]: ERROR nova.db.main.api [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 465.278768] env[62368]: result = function(*args, **kwargs) [ 465.278768] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 465.278768] env[62368]: return func(*args, **kwargs) [ 465.278768] env[62368]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 465.278768] env[62368]: result = fn(*args, **kwargs) [ 465.278768] env[62368]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 465.278768] env[62368]: return f(*args, **kwargs) [ 465.278768] env[62368]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 465.278768] env[62368]: return db.service_get_minimum_version(context, binaries) [ 465.278768] env[62368]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 465.278768] env[62368]: _check_db_access() [ 465.278768] env[62368]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 465.278768] env[62368]: stacktrace = ''.join(traceback.format_stack()) [ 465.278768] env[62368]: [ 465.279582] env[62368]: ERROR nova.db.main.api [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 465.279582] env[62368]: result = function(*args, **kwargs) [ 465.279582] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 465.279582] env[62368]: return func(*args, **kwargs) [ 465.279582] env[62368]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 465.279582] env[62368]: result = fn(*args, **kwargs) [ 465.279582] env[62368]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 465.279582] env[62368]: return f(*args, **kwargs) [ 465.279582] env[62368]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 465.279582] env[62368]: return db.service_get_minimum_version(context, binaries) [ 465.279582] env[62368]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 465.279582] env[62368]: _check_db_access() [ 465.279582] env[62368]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 465.279582] env[62368]: stacktrace = ''.join(traceback.format_stack()) [ 465.279582] env[62368]: [ 465.280022] env[62368]: WARNING nova.objects.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 465.280176] env[62368]: WARNING nova.objects.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Failed to get minimum service version for cell d117b194-c7ff-4d07-bcc2-1fe5b28965c2 [ 465.280623] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Acquiring lock "singleton_lock" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 465.280791] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Acquired lock "singleton_lock" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 465.281041] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Releasing lock "singleton_lock" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 465.281381] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Full set of CONF: {{(pid=62368) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 465.281531] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ******************************************************************************** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 465.281659] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Configuration options gathered from: {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 465.281798] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 465.281986] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 465.282124] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ================================================================================ {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 465.282337] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] allow_resize_to_same_host = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.282508] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] arq_binding_timeout = 300 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.282639] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] backdoor_port = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.282766] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] backdoor_socket = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.282927] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] block_device_allocate_retries = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.283097] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] block_device_allocate_retries_interval = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.283269] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cert = self.pem {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.283438] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.283609] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute_monitors = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.283779] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] config_dir = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.283953] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] config_drive_format = iso9660 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.284104] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.284275] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] config_source = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.284514] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] console_host = devstack {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.284709] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] control_exchange = nova {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.284872] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cpu_allocation_ratio = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.285048] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] daemon = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.285219] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] debug = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.285379] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] default_access_ip_network_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.285546] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] default_availability_zone = nova {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.285705] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] default_ephemeral_format = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.285864] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] default_green_pool_size = 1000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.286108] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.286277] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] default_schedule_zone = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.286440] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] disk_allocation_ratio = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.286599] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] enable_new_services = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.286778] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] enabled_apis = ['osapi_compute'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.286940] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] enabled_ssl_apis = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.287113] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] flat_injected = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.287277] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] force_config_drive = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.287438] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] force_raw_images = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.287604] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] graceful_shutdown_timeout = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.287768] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] heal_instance_info_cache_interval = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.287977] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] host = cpu-1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.288229] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.288426] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] initial_disk_allocation_ratio = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.288594] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] initial_ram_allocation_ratio = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.288811] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.288979] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] instance_build_timeout = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.289164] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] instance_delete_interval = 300 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.289335] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] instance_format = [instance: %(uuid)s] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.289504] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] instance_name_template = instance-%08x {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.289668] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] instance_usage_audit = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.289838] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] instance_usage_audit_period = month {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.290015] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.290221] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] instances_path = /opt/stack/data/nova/instances {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.290390] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] internal_service_availability_zone = internal {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.290556] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] key = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.290717] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] live_migration_retry_count = 30 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.290885] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_color = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.291058] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_config_append = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.291237] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.291405] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_dir = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.291568] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.291698] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_options = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.291862] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_rotate_interval = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.292043] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_rotate_interval_type = days {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.292216] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] log_rotation_type = none {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.292348] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.292474] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.292643] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.292809] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.292938] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.293115] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] long_rpc_timeout = 1800 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.293281] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] max_concurrent_builds = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.293446] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] max_concurrent_live_migrations = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.293605] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] max_concurrent_snapshots = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.293763] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] max_local_block_devices = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.293924] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] max_logfile_count = 30 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.294090] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] max_logfile_size_mb = 200 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.294254] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] maximum_instance_delete_attempts = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.294422] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] metadata_listen = 0.0.0.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.294589] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] metadata_listen_port = 8775 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.294757] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] metadata_workers = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.294915] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] migrate_max_retries = -1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.295092] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] mkisofs_cmd = genisoimage {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.295304] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] my_block_storage_ip = 10.180.1.21 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.295440] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] my_ip = 10.180.1.21 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.295605] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] network_allocate_retries = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.295783] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.295951] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] osapi_compute_listen = 0.0.0.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.296128] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] osapi_compute_listen_port = 8774 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.296298] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] osapi_compute_unique_server_name_scope = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.296492] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] osapi_compute_workers = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.296668] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] password_length = 12 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.296832] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] periodic_enable = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.296994] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] periodic_fuzzy_delay = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.297180] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] pointer_model = usbtablet {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.297349] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] preallocate_images = none {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.297512] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] publish_errors = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.297638] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] pybasedir = /opt/stack/nova {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.297792] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ram_allocation_ratio = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.297948] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] rate_limit_burst = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.298127] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] rate_limit_except_level = CRITICAL {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.298290] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] rate_limit_interval = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.298450] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] reboot_timeout = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.298606] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] reclaim_instance_interval = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.298761] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] record = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.298926] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] reimage_timeout_per_gb = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.299099] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] report_interval = 120 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.299266] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] rescue_timeout = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.299427] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] reserved_host_cpus = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.299585] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] reserved_host_disk_mb = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.299805] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] reserved_host_memory_mb = 512 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.299981] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] reserved_huge_pages = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.300173] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] resize_confirm_window = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.300346] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] resize_fs_using_block_device = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.300511] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] resume_guests_state_on_host_boot = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.300679] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.300843] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] rpc_response_timeout = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.301014] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] run_external_periodic_tasks = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.301192] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] running_deleted_instance_action = reap {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.301357] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] running_deleted_instance_poll_interval = 1800 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.301517] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] running_deleted_instance_timeout = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.301680] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler_instance_sync_interval = 120 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.301843] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_down_time = 720 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.302015] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] servicegroup_driver = db {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.302177] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] shell_completion = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.302340] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] shelved_offload_time = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.302496] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] shelved_poll_interval = 3600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.302662] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] shutdown_timeout = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.302824] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] source_is_ipv6 = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.302983] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ssl_only = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.303242] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.303412] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] sync_power_state_interval = 600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.303576] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] sync_power_state_pool_size = 1000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.303743] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] syslog_log_facility = LOG_USER {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.303901] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] tempdir = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.304071] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] timeout_nbd = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.304240] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] transport_url = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.304401] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] update_resources_interval = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.304560] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] use_cow_images = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.304718] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] use_eventlog = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.304874] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] use_journal = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.305040] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] use_json = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.305201] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] use_rootwrap_daemon = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.305358] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] use_stderr = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.305517] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] use_syslog = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.305673] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vcpu_pin_set = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.305836] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plugging_is_fatal = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.306007] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plugging_timeout = 300 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.306173] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] virt_mkfs = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.306335] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] volume_usage_poll_interval = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.306493] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] watch_log_file = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.306658] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] web = /usr/share/spice-html5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 465.306838] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_concurrency.disable_process_locking = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.307392] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.307591] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.307766] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.307943] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.308152] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.308340] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.308531] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.auth_strategy = keystone {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.308704] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.compute_link_prefix = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.308882] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.309070] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.dhcp_domain = novalocal {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.309247] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.enable_instance_password = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.309413] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.glance_link_prefix = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.309579] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.309751] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.309920] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.instance_list_per_project_cells = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.310092] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.list_records_by_skipping_down_cells = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.310287] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.local_metadata_per_cell = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.310469] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.max_limit = 1000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.310640] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.metadata_cache_expiration = 15 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.310814] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.neutron_default_tenant_id = default {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.310985] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.response_validation = warn {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.311178] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.use_neutron_default_nets = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.311367] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.311533] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.311707] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.311879] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.312063] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.vendordata_dynamic_targets = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.312232] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.vendordata_jsonfile_path = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.312417] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.312609] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.backend = dogpile.cache.memcached {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.312780] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.backend_argument = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.312954] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.config_prefix = cache.oslo {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.313142] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.dead_timeout = 60.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.313310] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.debug_cache_backend = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.313474] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.enable_retry_client = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.313640] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.enable_socket_keepalive = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.313810] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.enabled = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.313976] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.enforce_fips_mode = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.314155] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.expiration_time = 600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.314317] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.hashclient_retry_attempts = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.314487] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.hashclient_retry_delay = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.314647] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_dead_retry = 300 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.314807] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_password = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.314970] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.315144] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.315306] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_pool_maxsize = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.315469] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.315630] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_sasl_enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.315808] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.315978] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_socket_timeout = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.316151] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.memcache_username = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.316320] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.proxies = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.316485] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.redis_db = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.316646] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.redis_password = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.316817] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.redis_sentinel_service_name = mymaster {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.316991] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.317178] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.redis_server = localhost:6379 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.317349] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.redis_socket_timeout = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.317508] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.redis_username = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.317671] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.retry_attempts = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.317837] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.retry_delay = 0.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.318007] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.socket_keepalive_count = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.318205] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.socket_keepalive_idle = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.318380] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.socket_keepalive_interval = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.318542] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.tls_allowed_ciphers = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.318704] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.tls_cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.318861] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.tls_certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.319033] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.tls_enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.319197] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cache.tls_keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.319368] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.319541] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.auth_type = password {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.319703] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.319877] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.catalog_info = volumev3::publicURL {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.320049] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.320244] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.320423] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.cross_az_attach = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.320589] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.debug = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.320753] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.endpoint_template = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.320919] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.http_retries = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.321128] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.321327] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.321508] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.os_region_name = RegionOne {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.321675] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.321836] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cinder.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.322015] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.322188] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.cpu_dedicated_set = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.322351] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.cpu_shared_set = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.322517] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.image_type_exclude_list = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.322683] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.322845] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.max_concurrent_disk_ops = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.323014] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.max_disk_devices_to_attach = -1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.323179] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.323350] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.323513] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.resource_provider_association_refresh = 300 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.323676] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.323838] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.shutdown_retry_interval = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.324022] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.324205] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] conductor.workers = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.324387] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] console.allowed_origins = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.324548] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] console.ssl_ciphers = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.324721] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] console.ssl_minimum_version = default {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.324893] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] consoleauth.enforce_session_timeout = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.325075] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] consoleauth.token_ttl = 600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.325244] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.325406] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.325568] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.325728] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.connect_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.325885] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.connect_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.326052] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.endpoint_override = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.326217] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.326376] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.326535] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.max_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.326693] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.min_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.326849] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.region_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.327014] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.retriable_status_codes = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.327182] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.service_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.327352] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.service_type = accelerator {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.327513] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.327671] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.status_code_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.327828] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.status_code_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.327985] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.328204] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.328378] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] cyborg.version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.328564] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.backend = sqlalchemy {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.328730] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.connection = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.328898] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.connection_debug = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.329078] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.connection_parameters = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.329249] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.connection_recycle_time = 3600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.329415] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.connection_trace = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.329575] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.db_inc_retry_interval = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.329738] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.db_max_retries = 20 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.329899] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.db_max_retry_interval = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.330070] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.db_retry_interval = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.330284] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.max_overflow = 50 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.330463] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.max_pool_size = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.330628] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.max_retries = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.330800] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.330961] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.mysql_wsrep_sync_wait = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.331158] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.pool_timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.331361] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.retry_interval = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.331529] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.slave_connection = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.331694] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.sqlite_synchronous = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.331856] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] database.use_db_reconnect = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.332046] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.backend = sqlalchemy {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.332221] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.connection = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.332390] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.connection_debug = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.332561] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.connection_parameters = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.332729] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.connection_recycle_time = 3600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.332890] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.connection_trace = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.333063] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.db_inc_retry_interval = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.333231] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.db_max_retries = 20 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.333446] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.db_max_retry_interval = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.333647] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.db_retry_interval = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.333818] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.max_overflow = 50 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.333984] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.max_pool_size = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.334163] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.max_retries = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.334334] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.334497] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.334658] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.pool_timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.334820] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.retry_interval = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.334980] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.slave_connection = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.335157] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] api_database.sqlite_synchronous = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.335335] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] devices.enabled_mdev_types = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.335517] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.335689] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ephemeral_storage_encryption.default_format = luks {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.335855] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ephemeral_storage_encryption.enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.336029] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.336203] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.api_servers = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.336372] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.336536] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.336702] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.336865] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.connect_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.337032] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.connect_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.337202] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.debug = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.337372] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.default_trusted_certificate_ids = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.337534] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.enable_certificate_validation = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.337696] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.enable_rbd_download = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.337854] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.endpoint_override = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.338028] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.338220] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.338388] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.max_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.338547] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.min_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.338710] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.num_retries = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.338878] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.rbd_ceph_conf = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.339052] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.rbd_connect_timeout = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.339225] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.rbd_pool = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.339395] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.rbd_user = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.339557] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.region_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.339719] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.retriable_status_codes = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.339878] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.service_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.340057] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.service_type = image {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.340253] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.340423] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.status_code_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.340584] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.status_code_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.340744] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.340931] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.341114] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.verify_glance_signatures = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.341300] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] glance.version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.341478] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] guestfs.debug = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.341647] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] mks.enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.342010] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.342231] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] image_cache.manager_interval = 2400 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.342420] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] image_cache.precache_concurrency = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.342597] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] image_cache.remove_unused_base_images = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.342765] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.342934] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.343123] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] image_cache.subdirectory_name = _base {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.343303] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.api_max_retries = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.343469] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.api_retry_interval = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.343629] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.343790] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.auth_type = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.343948] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.344117] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.344281] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.344447] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.conductor_group = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.344605] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.connect_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.344763] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.connect_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.344920] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.endpoint_override = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.345093] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.345256] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.345418] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.max_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.345574] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.min_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.345738] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.peer_list = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.345897] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.region_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.346065] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.retriable_status_codes = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.346232] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.serial_console_state_timeout = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.346395] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.service_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.346566] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.service_type = baremetal {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.346733] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.shard = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.346893] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.347061] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.status_code_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.347223] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.status_code_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.347384] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.347562] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.347725] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ironic.version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.347904] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.348107] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] key_manager.fixed_key = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.348301] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.348469] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.barbican_api_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.348630] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.barbican_endpoint = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.348801] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.barbican_endpoint_type = public {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.348962] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.barbican_region_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.349137] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.349299] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.349465] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.349629] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.349785] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.349948] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.number_of_retries = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.350119] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.retry_delay = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.350305] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.send_service_user_token = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.350479] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.350638] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.350800] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.verify_ssl = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.350960] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican.verify_ssl_path = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.351137] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.351325] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.auth_type = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.351488] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.351646] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.351807] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.351969] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.352139] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.352304] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.352466] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] barbican_service_user.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.352630] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.approle_role_id = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.352789] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.approle_secret_id = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.352959] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.kv_mountpoint = secret {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.353131] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.kv_path = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.353306] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.kv_version = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.353528] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.namespace = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.353703] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.root_token_id = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.353865] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.ssl_ca_crt_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.354044] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.timeout = 60.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.354216] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.use_ssl = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.354393] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.354563] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.354730] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.auth_type = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.354889] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.355059] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.355229] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.355389] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.connect_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.355548] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.connect_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.355705] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.endpoint_override = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.355866] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.356033] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.356195] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.max_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.356356] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.min_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.356567] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.region_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.356756] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.retriable_status_codes = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.356919] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.service_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.357103] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.service_type = identity {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.357271] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.357434] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.status_code_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.357594] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.status_code_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.357753] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.357934] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.358107] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] keystone.version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.358310] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.connection_uri = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.358478] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.cpu_mode = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.358646] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.cpu_model_extra_flags = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.358814] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.cpu_models = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.358985] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.cpu_power_governor_high = performance {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.359166] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.cpu_power_governor_low = powersave {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.359332] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.cpu_power_management = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.359502] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.359665] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.device_detach_attempts = 8 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.359827] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.device_detach_timeout = 20 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.359990] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.disk_cachemodes = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.360171] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.disk_prefix = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.360369] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.enabled_perf_events = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.360537] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.file_backed_memory = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.360706] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.gid_maps = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.360868] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.hw_disk_discard = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.361036] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.hw_machine_type = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.361230] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.images_rbd_ceph_conf = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.361411] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.361575] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.361748] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.images_rbd_glance_store_name = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.361923] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.images_rbd_pool = rbd {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.362108] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.images_type = default {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.362278] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.images_volume_group = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.362442] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.inject_key = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.362607] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.inject_partition = -2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.362768] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.inject_password = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.362954] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.iscsi_iface = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.363149] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.iser_use_multipath = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.363397] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_bandwidth = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.363632] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.363810] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_downtime = 500 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.363979] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.364161] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.364329] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_inbound_addr = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.364497] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.364662] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_permit_post_copy = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.364825] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_scheme = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.364999] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_timeout_action = abort {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.365181] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_tunnelled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.365348] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_uri = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.365513] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.live_migration_with_native_tls = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.365676] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.max_queues = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.365842] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.366083] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.366252] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.nfs_mount_options = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.366546] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.366719] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.366884] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.num_iser_scan_tries = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.367057] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.num_memory_encrypted_guests = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.367227] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.367393] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.num_pcie_ports = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.367561] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.num_volume_scan_tries = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.367730] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.pmem_namespaces = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.367893] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.quobyte_client_cfg = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.368205] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.368390] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rbd_connect_timeout = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.368561] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.368725] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.368886] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rbd_secret_uuid = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.369057] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rbd_user = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.369228] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.369402] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.remote_filesystem_transport = ssh {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.369563] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rescue_image_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.369723] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rescue_kernel_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.369912] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rescue_ramdisk_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.370094] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.370283] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.rx_queue_size = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.370467] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.smbfs_mount_options = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.370747] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.370918] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.snapshot_compression = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.371093] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.snapshot_image_format = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.371343] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.371518] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.sparse_logical_volumes = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.371683] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.swtpm_enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.371853] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.swtpm_group = tss {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.372033] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.swtpm_user = tss {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.372210] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.sysinfo_serial = unique {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.372378] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.tb_cache_size = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.372538] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.tx_queue_size = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.372706] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.uid_maps = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.372869] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.use_virtio_for_bridges = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.373048] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.virt_type = kvm {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.373224] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.volume_clear = zero {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.373392] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.volume_clear_size = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.373557] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.volume_use_multipath = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.373717] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.vzstorage_cache_path = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.373889] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.374137] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.vzstorage_mount_group = qemu {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.374319] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.vzstorage_mount_opts = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.374491] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.374768] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.374944] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.vzstorage_mount_user = stack {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.375120] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.375301] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.375479] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.auth_type = password {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.375669] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.375834] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.375996] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.376172] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.connect_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.376332] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.connect_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.376499] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.default_floating_pool = public {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.376656] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.endpoint_override = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.376816] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.extension_sync_interval = 600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.376975] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.http_retries = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.377144] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.377302] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.377461] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.max_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.377633] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.377789] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.min_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.377952] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.ovs_bridge = br-int {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.378143] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.physnets = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.378323] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.region_name = RegionOne {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.378489] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.retriable_status_codes = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.378655] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.service_metadata_proxy = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.378812] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.service_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.378978] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.service_type = network {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.379148] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.379307] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.status_code_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.379465] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.status_code_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.379618] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.379795] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.379985] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] neutron.version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.380171] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] notifications.bdms_in_notifications = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.380363] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] notifications.default_level = INFO {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.380540] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] notifications.notification_format = unversioned {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.380703] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] notifications.notify_on_state_change = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.380876] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.381058] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] pci.alias = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.381247] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] pci.device_spec = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.381420] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] pci.report_in_placement = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.381593] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.381766] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.auth_type = password {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.381933] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.382105] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.382268] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.382433] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.382590] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.connect_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.382750] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.connect_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.382908] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.default_domain_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.383073] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.default_domain_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.383235] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.domain_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.383394] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.domain_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.383546] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.endpoint_override = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.383703] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.383859] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.384024] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.max_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.384178] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.min_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.384346] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.password = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.384501] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.project_domain_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.384668] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.project_domain_name = Default {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.384837] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.project_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.385010] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.project_name = service {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.385184] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.region_name = RegionOne {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.385379] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.retriable_status_codes = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.385645] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.service_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.385934] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.service_type = placement {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.386237] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.386517] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.status_code_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.386778] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.status_code_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.386972] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.system_scope = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.387153] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.387319] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.trust_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.387485] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.user_domain_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.387700] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.user_domain_name = Default {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.387874] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.user_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.388060] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.username = nova {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.388248] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.388414] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] placement.version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.388592] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.cores = 20 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.388759] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.count_usage_from_placement = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.388933] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.389121] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.injected_file_content_bytes = 10240 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.389292] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.injected_file_path_length = 255 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.389463] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.injected_files = 5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.389627] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.instances = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.389792] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.key_pairs = 100 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.389956] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.metadata_items = 128 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.390132] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.ram = 51200 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.390325] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.recheck_quota = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.390498] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.server_group_members = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.390712] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] quota.server_groups = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.391016] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.391312] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.391603] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.image_metadata_prefilter = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.391883] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.392127] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.max_attempts = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.392310] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.max_placement_results = 1000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.392510] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.392684] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.query_placement_for_image_type_support = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.392850] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.393032] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] scheduler.workers = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.393215] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.393387] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.393565] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.393732] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.393898] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.394074] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.394246] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.394445] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.394617] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.host_subset_size = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.394778] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.394938] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.395115] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.395285] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.isolated_hosts = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.395455] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.isolated_images = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.395618] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.395783] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.395953] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.396129] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.pci_in_placement = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.396295] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.396461] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.396625] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.396788] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.396954] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.397131] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.397299] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.track_instance_changes = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.397481] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.397652] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] metrics.required = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.397815] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] metrics.weight_multiplier = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.397976] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.398178] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] metrics.weight_setting = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.398511] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.398712] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] serial_console.enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.398900] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] serial_console.port_range = 10000:20000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.399084] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.399258] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.399428] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] serial_console.serialproxy_port = 6083 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.399597] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.399767] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.auth_type = password {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.399926] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.400096] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.400280] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.400448] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.400605] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.400820] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.send_service_user_token = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.401026] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.401197] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] service_user.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.401367] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.agent_enabled = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.401530] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.401853] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.402052] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.402226] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.html5proxy_port = 6082 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.402390] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.image_compression = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.402575] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.jpeg_compression = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.402739] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.playback_compression = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.402903] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.require_secure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.403081] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.server_listen = 127.0.0.1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.403255] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.403416] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.streaming_mode = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.403571] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] spice.zlib_compression = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.403736] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] upgrade_levels.baseapi = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.403904] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] upgrade_levels.compute = auto {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.404073] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] upgrade_levels.conductor = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.404235] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] upgrade_levels.scheduler = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.404403] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.404562] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.auth_type = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.404720] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.404876] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.405045] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.405208] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.405368] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.405525] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.405680] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vendordata_dynamic_auth.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.405851] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.api_retry_count = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.406016] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.ca_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.406192] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.cache_prefix = devstack-image-cache {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.406359] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.cluster_name = testcl1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.406521] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.connection_pool_size = 10 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.406680] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.console_delay_seconds = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.406846] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.datastore_regex = ^datastore.* {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.407059] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.407233] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.host_password = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.407400] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.host_port = 443 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.407563] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.host_username = administrator@vsphere.local {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.407727] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.insecure = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.407887] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.integration_bridge = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.408060] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.maximum_objects = 100 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.408244] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.pbm_default_policy = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.408414] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.pbm_enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.408569] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.pbm_wsdl_location = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.408735] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.408893] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.serial_port_proxy_uri = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.409062] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.serial_port_service_uri = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.409231] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.task_poll_interval = 0.5 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.409403] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.use_linked_clone = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.409569] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.vnc_keymap = en-us {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.409732] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.vnc_port = 5900 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.409895] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vmware.vnc_port_total = 10000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.410087] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.auth_schemes = ['none'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.410312] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.410627] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.410903] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.411113] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.novncproxy_port = 6080 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.411304] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.server_listen = 127.0.0.1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.411478] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.411641] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.vencrypt_ca_certs = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.411801] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.vencrypt_client_cert = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.411958] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vnc.vencrypt_client_key = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.412148] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.412312] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.disable_deep_image_inspection = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.412474] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.412635] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.412796] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.412957] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.disable_rootwrap = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.413132] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.enable_numa_live_migration = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.413295] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.413458] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.413619] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.413779] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.libvirt_disable_apic = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.413936] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.414108] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.414271] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.414435] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.414594] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.414753] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.414914] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.415084] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.415249] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.415417] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.415601] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.415772] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.client_socket_timeout = 900 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.415937] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.default_pool_size = 1000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.416113] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.keep_alive = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.416282] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.max_header_line = 16384 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.416446] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.secure_proxy_ssl_header = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.416605] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.ssl_ca_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.416774] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.ssl_cert_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.416935] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.ssl_key_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.417109] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.tcp_keepidle = 600 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.417297] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.417464] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] zvm.ca_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.417625] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] zvm.cloud_connector_url = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.417904] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.418100] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] zvm.reachable_timeout = 300 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.418299] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.enforce_new_defaults = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.418694] env[62368]: WARNING oslo_config.cfg [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 465.418879] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.enforce_scope = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.419071] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.policy_default_rule = default {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.419258] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.419438] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.policy_file = policy.yaml {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.419617] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.419780] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.419943] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.420118] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.420307] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.420485] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.420662] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.420841] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.connection_string = messaging:// {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.421045] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.enabled = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.421238] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.es_doc_type = notification {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.421406] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.es_scroll_size = 10000 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.422625] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.es_scroll_time = 2m {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.422625] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.filter_error_trace = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.422625] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.hmac_keys = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.422625] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.sentinel_service_name = mymaster {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.422625] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.socket_timeout = 0.1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.422625] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.trace_requests = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.422833] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler.trace_sqlalchemy = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.422864] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler_jaeger.process_tags = {} {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.423033] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler_jaeger.service_name_prefix = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.423197] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] profiler_otlp.service_name_prefix = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427019] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] remote_debug.host = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427019] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] remote_debug.port = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427019] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427019] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427019] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427019] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427019] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427293] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427293] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427293] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427293] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427293] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427293] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427293] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427482] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427482] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427482] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427482] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427482] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427482] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427482] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427661] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427661] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427661] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427661] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427661] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427789] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.427982] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.428051] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.428189] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.428367] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.428535] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.ssl = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.428708] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.428877] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.429052] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.429228] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.429399] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.ssl_version = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.429561] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.429747] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.429918] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_notifications.retry = -1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.430116] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.430321] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_messaging_notifications.transport_url = **** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.430500] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.auth_section = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.430664] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.auth_type = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.430826] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.cafile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.430982] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.certfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.431183] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.collect_timing = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.431349] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.connect_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.431509] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.connect_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.431668] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.endpoint_id = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.431827] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.endpoint_override = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.431987] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.insecure = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.432157] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.keyfile = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.432317] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.max_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.432475] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.min_version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.432633] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.region_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.432793] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.retriable_status_codes = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.432950] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.service_name = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.433127] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.service_type = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.433291] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.split_loggers = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.433453] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.status_code_retries = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.433613] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.status_code_retry_delay = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.433770] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.timeout = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.433929] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.valid_interfaces = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.434098] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_limit.version = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.434325] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_reports.file_event_handler = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.434518] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.434682] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] oslo_reports.log_dir = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.434855] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.435031] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.435198] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.435370] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.435535] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.435697] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.435870] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.436040] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_ovs_privileged.group = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.436207] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.436377] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.436540] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.436698] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] vif_plug_ovs_privileged.user = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.436869] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_linux_bridge.flat_interface = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.437062] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.437241] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.437418] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.437589] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.437757] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.437923] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.438114] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.438318] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.438495] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_ovs.isolate_vif = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.438666] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.438832] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.439009] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.439189] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_ovs.ovsdb_interface = native {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.439354] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_vif_ovs.per_port_bridge = False {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.439520] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_brick.lock_path = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.439684] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.439846] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.440025] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] privsep_osbrick.capabilities = [21] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.440206] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] privsep_osbrick.group = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.440384] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] privsep_osbrick.helper_command = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.440556] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.440720] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.440881] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] privsep_osbrick.user = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.441065] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.441299] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] nova_sys_admin.group = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.441436] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] nova_sys_admin.helper_command = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.441604] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.441767] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.441925] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] nova_sys_admin.user = None {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 465.442066] env[62368]: DEBUG oslo_service.service [None req-dcede2b1-d6d5-45f0-815d-910de9dcdb6c None None] ******************************************************************************** {{(pid=62368) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 465.442566] env[62368]: INFO nova.service [-] Starting compute node (version 30.0.1) [ 465.946066] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Getting list of instances from cluster (obj){ [ 465.946066] env[62368]: value = "domain-c8" [ 465.946066] env[62368]: _type = "ClusterComputeResource" [ 465.946066] env[62368]: } {{(pid=62368) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 465.947212] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79489d00-9e4e-4773-b292-03ff32e288a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.956868] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Got total of 0 instances {{(pid=62368) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 465.957412] env[62368]: WARNING nova.virt.vmwareapi.driver [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 465.957879] env[62368]: INFO nova.virt.node [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Generated node identity 2202a74c-753d-4e1d-a031-7cefe24ee9d6 [ 465.958132] env[62368]: INFO nova.virt.node [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Wrote node identity 2202a74c-753d-4e1d-a031-7cefe24ee9d6 to /opt/stack/data/n-cpu-1/compute_id [ 466.461171] env[62368]: WARNING nova.compute.manager [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Compute nodes ['2202a74c-753d-4e1d-a031-7cefe24ee9d6'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 467.468057] env[62368]: INFO nova.compute.manager [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 468.473479] env[62368]: WARNING nova.compute.manager [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 468.473816] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 468.473917] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 468.474287] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 468.474588] env[62368]: DEBUG nova.compute.resource_tracker [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 468.475651] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2401bbef-c611-4df9-8590-1e6f9806d5ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.487020] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5e9192-b9cc-405a-813e-2f56c52477bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.497146] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d76b5a-8a4d-4ecb-8284-c3b39b0fb96e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.503539] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e55a29e-7405-42ea-9911-d0e134204d67 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.533376] env[62368]: DEBUG nova.compute.resource_tracker [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181439MB free_disk=157GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 468.533723] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 468.534033] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 469.037864] env[62368]: WARNING nova.compute.resource_tracker [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] No compute node record for cpu-1:2202a74c-753d-4e1d-a031-7cefe24ee9d6: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 2202a74c-753d-4e1d-a031-7cefe24ee9d6 could not be found. [ 469.540983] env[62368]: INFO nova.compute.resource_tracker [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 [ 471.051472] env[62368]: DEBUG nova.compute.resource_tracker [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 471.051472] env[62368]: DEBUG nova.compute.resource_tracker [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 471.215958] env[62368]: INFO nova.scheduler.client.report [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] [req-27fdd3fb-b5a2-436c-a2a3-8820b0f797ef] Created resource provider record via placement API for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 471.356146] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b780df3-6e93-447d-a405-449b7d8906cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 471.364387] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350f2166-1534-4ac8-91cf-d72ad056245c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 471.396442] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f08ca0-6a85-4d49-92e6-af99dc8611bc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 471.402022] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c08f87d-a9d4-4d72-aa7a-257ed0e5b881 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 471.413897] env[62368]: DEBUG nova.compute.provider_tree [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 471.954014] env[62368]: DEBUG nova.scheduler.client.report [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 471.954014] env[62368]: DEBUG nova.compute.provider_tree [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 0 to 1 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 471.954014] env[62368]: DEBUG nova.compute.provider_tree [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 472.005015] env[62368]: DEBUG nova.compute.provider_tree [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 1 to 2 during operation: update_traits {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 472.510251] env[62368]: DEBUG nova.compute.resource_tracker [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 472.510251] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.974s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 472.510251] env[62368]: DEBUG nova.service [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Creating RPC server for service compute {{(pid=62368) start /opt/stack/nova/nova/service.py:186}} [ 472.521910] env[62368]: DEBUG nova.service [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] Join ServiceGroup membership for this service compute {{(pid=62368) start /opt/stack/nova/nova/service.py:203}} [ 472.521910] env[62368]: DEBUG nova.servicegroup.drivers.db [None req-b9a129b9-e23e-47e7-abe9-3e034e993a1f None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62368) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 509.613820] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "5c652903-a025-4530-96ee-428da51c8b71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.613820] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "5c652903-a025-4530-96ee-428da51c8b71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.117811] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 510.618432] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "0eb2a772-7635-481a-ae2c-9f23669dd175" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.618813] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "0eb2a772-7635-481a-ae2c-9f23669dd175" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.659276] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.659547] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.661452] env[62368]: INFO nova.compute.claims [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 510.944268] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "c198d320-9b6a-4177-9e5e-1d577f9a91ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.944268] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "c198d320-9b6a-4177-9e5e-1d577f9a91ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.122173] env[62368]: DEBUG nova.compute.manager [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.274461] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Acquiring lock "0144a1d0-bc85-4c8c-9342-8c12da4b8967" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.274461] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Lock "0144a1d0-bc85-4c8c-9342-8c12da4b8967" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.447750] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.654758] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.776476] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.789682] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405dd14b-15da-4a42-a869-6db4da708252 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.803569] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf7813e-8b96-43e8-ace8-369c325c8d79 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.843799] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981cb09a-3cdb-4ff9-8f65-6cb99617f164 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.853322] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68918ee-44f8-4dc9-a3fc-a091dfec1137 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.870168] env[62368]: DEBUG nova.compute.provider_tree [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 511.983984] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.102747] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Acquiring lock "50feff45-9ede-4d4d-b136-bfb5fd916dc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.102972] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Lock "50feff45-9ede-4d4d-b136-bfb5fd916dc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.306606] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.376899] env[62368]: DEBUG nova.scheduler.client.report [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 512.523539] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 512.605735] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 512.887814] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.889325] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 512.895891] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.241s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.897323] env[62368]: INFO nova.compute.claims [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.032288] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Getting list of instances from cluster (obj){ [ 513.032288] env[62368]: value = "domain-c8" [ 513.032288] env[62368]: _type = "ClusterComputeResource" [ 513.032288] env[62368]: } {{(pid=62368) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 513.033351] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b05d62-a41e-408e-9caf-701d04c8e20e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.050346] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Got total of 0 instances {{(pid=62368) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 513.051817] env[62368]: WARNING nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] While synchronizing instance power states, found 1 instances in the database and 0 instances on the hypervisor. [ 513.051817] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Triggering sync for uuid 5c652903-a025-4530-96ee-428da51c8b71 {{(pid=62368) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 513.052235] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "5c652903-a025-4530-96ee-428da51c8b71" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.052497] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 513.052894] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Getting list of instances from cluster (obj){ [ 513.052894] env[62368]: value = "domain-c8" [ 513.052894] env[62368]: _type = "ClusterComputeResource" [ 513.052894] env[62368]: } {{(pid=62368) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 513.053848] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30edf57b-fffa-4310-b62d-2a8215adc76b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.063141] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Got total of 0 instances {{(pid=62368) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 513.132232] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.402785] env[62368]: DEBUG nova.compute.utils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.404991] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 513.405255] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 513.911984] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 514.068835] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10344556-b1bd-43ee-8d8d-869484ced3c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.078460] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa6a379-70d1-49d4-b0ca-77322d28f166 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.124884] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b26d5f1-6b9b-464d-aa7e-1a4b1920ded2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.133884] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d15245a-a8f3-41c2-bab0-6728e177f46e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.153605] env[62368]: DEBUG nova.compute.provider_tree [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 514.499149] env[62368]: DEBUG nova.policy [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f3afe492e7e47f28c3c8dad61f819d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81b69987192149ddb6dfcafc1f7770fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 514.661385] env[62368]: DEBUG nova.scheduler.client.report [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 514.931131] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 514.966659] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 514.967400] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 514.967400] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 514.967400] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 514.967400] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 514.967812] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 514.967812] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 514.968078] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 514.968475] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 514.968747] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 514.970089] env[62368]: DEBUG nova.virt.hardware [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 514.971164] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84598c5d-ac85-43f7-8c5c-ebf822b892b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.981117] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38c94c8-3ab4-4a51-8ec9-4dce8f825f8f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.998965] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21dff70-af17-448d-b25c-c9f1c2a2fb97 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.167709] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.272s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.168808] env[62368]: DEBUG nova.compute.manager [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 515.172730] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.189s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.175594] env[62368]: INFO nova.compute.claims [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 515.679998] env[62368]: DEBUG nova.compute.utils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 515.681555] env[62368]: DEBUG nova.compute.manager [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Not allocating networking since 'none' was specified. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 516.040917] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Successfully created port: 72ed829c-2b51-4d83-87e0-ba3316d3d5ef {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 516.185179] env[62368]: DEBUG nova.compute.manager [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 516.291974] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8f8c3c-5856-4d84-b359-c20e899c30ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.305308] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f2d03d-44d4-4fc3-bd1e-f743d2e42f08 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.337987] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7854edad-af0f-4245-a6ae-8a220c81e118 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.347458] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06bdb77-5a57-4647-9e82-2247cdcdb3ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.363341] env[62368]: DEBUG nova.compute.provider_tree [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 516.866541] env[62368]: DEBUG nova.scheduler.client.report [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 517.203334] env[62368]: DEBUG nova.compute.manager [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 517.236957] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 517.236957] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 517.237108] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 517.237563] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 517.237563] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 517.237563] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 517.239719] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 517.240043] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 517.240357] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 517.240467] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 517.240811] env[62368]: DEBUG nova.virt.hardware [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 517.241765] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d376955b-4aa2-45e4-8bb9-b515db4445dd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.255374] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d498e344-3f44-427b-9903-1b219d7a6b16 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.273646] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 517.285859] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 517.288453] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f613fa59-88b5-4a39-bbde-c187e9c5d5d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.304110] env[62368]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 517.304110] env[62368]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62368) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 517.305172] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 517.305172] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating folder: Project (7cad8d0b574e44e797a2fb803765ac8b). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 517.305342] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-190b195d-8687-47fb-850f-b70043f1cc77 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.319179] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Created folder: Project (7cad8d0b574e44e797a2fb803765ac8b) in parent group-v259706. [ 517.319455] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating folder: Instances. Parent ref: group-v259710. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 517.319771] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d13a35a-dd64-4899-bad8-02c77de005d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.333908] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Created folder: Instances in parent group-v259710. [ 517.334225] env[62368]: DEBUG oslo.service.loopingcall [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 517.334460] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 517.334810] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b6adecd-edda-4a4b-b115-3383f244fe29 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.355611] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 517.355611] env[62368]: value = "task-1198121" [ 517.355611] env[62368]: _type = "Task" [ 517.355611] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.366177] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198121, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.374952] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.202s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.375376] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.378805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.072s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.381855] env[62368]: INFO nova.compute.claims [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 517.866519] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198121, 'name': CreateVM_Task, 'duration_secs': 0.317671} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.866801] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 517.867717] env[62368]: DEBUG oslo_vmware.service [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba66c31-fbc7-40e3-b86b-32c6802f9764 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.874977] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 517.874977] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.875936] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 517.875936] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc441f7a-b2de-4585-8768-f1261ff147f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.881202] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 517.881202] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]525273e6-9a36-6761-14ff-47f5a7f31fda" [ 517.881202] env[62368]: _type = "Task" [ 517.881202] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.885910] env[62368]: DEBUG nova.compute.utils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 517.895333] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 517.896259] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 517.906940] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 517.907849] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 517.907849] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 517.907849] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.909802] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 517.910345] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2210708e-e930-4f7f-805a-a419068d0f01 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.930495] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 517.930783] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 517.931613] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e9939d-ead0-4975-b747-e45d7ef91b07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.943771] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05c225a1-4f2f-4f77-9c79-3d5b5c69233c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.953601] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 517.953601] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5221911f-b877-6ea6-5974-b9cbd386b02a" [ 517.953601] env[62368]: _type = "Task" [ 517.953601] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.966541] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5221911f-b877-6ea6-5974-b9cbd386b02a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.071901] env[62368]: DEBUG nova.policy [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f53480e7585471fbf084ba6603d4ab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bdba0baa56ae465ca0045f41da157304', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 518.395835] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 518.473244] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Preparing fetch location {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 518.474779] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating directory with path [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 518.474779] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-094d76e5-82d9-438f-ad2b-45d6a0744166 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.497228] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Created directory with path [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 518.497424] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Fetch image to [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 518.497590] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Downloading image file data d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d to [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk on the data store datastore1 {{(pid=62368) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 518.498566] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a3b9b6-7b0b-49c6-8345-b4ec5465519b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.510520] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb06f8d2-497b-4f3e-bb94-c861a52655c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.518775] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4194cd97-d443-40b5-adc2-9d3caa57cea4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.531540] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819c1c47-be1a-40c3-b091-c5ad4d371694 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.573723] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1bad16-34fb-402b-8e8e-f3ef66ce7369 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.579634] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6844662-da55-48fd-8bab-24453e461a0a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.613477] env[62368]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-2d4806f4-9c74-490f-8a55-3b6668dcf054 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.618616] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88c9c9c-e6ea-42cf-809e-bb5f17b76306 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.627125] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9448c256-6e23-4dfd-83c3-d1c8933861b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.644469] env[62368]: DEBUG nova.compute.provider_tree [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.646506] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Downloading image file data d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d to the data store datastore1 {{(pid=62368) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 518.716915] env[62368]: DEBUG oslo_vmware.rw_handles [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62368) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 519.035348] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Acquiring lock "5d0e30b7-17b1-46c1-9cd0-05df00c9923e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.035348] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Lock "5d0e30b7-17b1-46c1-9cd0-05df00c9923e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.153186] env[62368]: DEBUG nova.scheduler.client.report [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.368431] env[62368]: DEBUG oslo_vmware.rw_handles [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Completed reading data from the image iterator. {{(pid=62368) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 519.368562] env[62368]: DEBUG oslo_vmware.rw_handles [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 519.408242] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 519.437992] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 519.438648] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 519.438648] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 519.439102] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 519.440060] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 519.440060] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 519.440060] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 519.440060] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 519.440236] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 519.440458] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 519.440642] env[62368]: DEBUG nova.virt.hardware [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 519.442812] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac49b5fe-4f97-41fb-b33a-8804db5c6f3b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.452745] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0b335f-d177-4b69-9a59-b4ce92d75e64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.515353] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Downloaded image file data d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d to vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk on the data store datastore1 {{(pid=62368) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 519.520536] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Caching image {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 519.520536] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Copying Virtual Disk [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk to [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 519.520997] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-295bf24b-f7a9-48a4-9f7b-30f5be5f22ee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.532131] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 519.532131] env[62368]: value = "task-1198123" [ 519.532131] env[62368]: _type = "Task" [ 519.532131] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.542027] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.545777] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.663314] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.284s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.663995] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 519.666845] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.536s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.673165] env[62368]: INFO nova.compute.claims [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.733183] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Successfully created port: f4814bcc-8ada-4a61-82a4-5dd21dd2ca86 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 520.061452] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198123, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.081525] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.186912] env[62368]: DEBUG nova.compute.utils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 520.190724] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 520.190724] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 520.314227] env[62368]: DEBUG nova.policy [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ac83b255b0c4565a8975960a39daeb9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb3861951fe84707b962058b871a25bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 520.353706] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Acquiring lock "431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.354744] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Lock "431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.515247] env[62368]: ERROR nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. [ 520.515247] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 520.515247] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.515247] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 520.515247] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 520.515247] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 520.515247] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 520.515247] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 520.515247] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.515247] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 520.515247] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.515247] env[62368]: ERROR nova.compute.manager raise self.value [ 520.515247] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 520.515247] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 520.515247] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.515247] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 520.515912] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.515912] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 520.515912] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. [ 520.515912] env[62368]: ERROR nova.compute.manager [ 520.515912] env[62368]: Traceback (most recent call last): [ 520.515912] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 520.515912] env[62368]: listener.cb(fileno) [ 520.515912] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.515912] env[62368]: result = function(*args, **kwargs) [ 520.515912] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.515912] env[62368]: return func(*args, **kwargs) [ 520.515912] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 520.515912] env[62368]: raise e [ 520.515912] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.515912] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 520.515912] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 520.515912] env[62368]: created_port_ids = self._update_ports_for_instance( [ 520.515912] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 520.515912] env[62368]: with excutils.save_and_reraise_exception(): [ 520.515912] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.515912] env[62368]: self.force_reraise() [ 520.515912] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.515912] env[62368]: raise self.value [ 520.515912] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 520.515912] env[62368]: updated_port = self._update_port( [ 520.515912] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.515912] env[62368]: _ensure_no_port_binding_failure(port) [ 520.515912] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.515912] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 520.516790] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. [ 520.516790] env[62368]: Removing descriptor: 15 [ 520.516790] env[62368]: ERROR nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] Traceback (most recent call last): [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] yield resources [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self.driver.spawn(context, instance, image_meta, [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 520.516790] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] vm_ref = self.build_virtual_machine(instance, [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] vif_infos = vmwarevif.get_vif_info(self._session, [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] for vif in network_info: [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] return self._sync_wrapper(fn, *args, **kwargs) [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self.wait() [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self[:] = self._gt.wait() [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] return self._exit_event.wait() [ 520.517118] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] result = hub.switch() [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] return self.greenlet.switch() [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] result = function(*args, **kwargs) [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] return func(*args, **kwargs) [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] raise e [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] nwinfo = self.network_api.allocate_for_instance( [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 520.517852] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] created_port_ids = self._update_ports_for_instance( [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] with excutils.save_and_reraise_exception(): [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self.force_reraise() [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] raise self.value [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] updated_port = self._update_port( [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] _ensure_no_port_binding_failure(port) [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.518261] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] raise exception.PortBindingFailed(port_id=port['id']) [ 520.518594] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] nova.exception.PortBindingFailed: Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. [ 520.518594] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] [ 520.518594] env[62368]: INFO nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Terminating instance [ 520.520022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.520193] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.520354] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 520.546841] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198123, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.695697} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.546841] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Copied Virtual Disk [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk to [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 520.549076] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleting the datastore file [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 520.550130] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-612c5c8b-c23b-4730-9b85-3c2138f531ab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.560100] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 520.560100] env[62368]: value = "task-1198124" [ 520.560100] env[62368]: _type = "Task" [ 520.560100] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.573116] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.695181] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 520.861397] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.865163] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b3fbb6-0c5a-44dd-bbc7-2103acffccdb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.872753] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 520.872753] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 520.872753] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 520.872753] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Rebuilding the list of instances to heal {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 520.879531] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900a67ab-169f-4758-a86a-2a398f572c61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.917188] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff6ec36-74f6-4b41-b9ac-353bedaf3f17 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.929150] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ab91c9-ed01-4649-9758-876d1990345f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.945872] env[62368]: DEBUG nova.compute.provider_tree [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.072576] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024324} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.072850] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 521.073171] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Moving file from [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d to [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d. {{(pid=62368) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 521.073414] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-9ac8e941-3a9e-4647-b2be-123a44f0c96b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.082578] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 521.082578] env[62368]: value = "task-1198125" [ 521.082578] env[62368]: _type = "Task" [ 521.082578] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.088184] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 521.095350] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198125, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.225685] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Successfully created port: 65fbf44f-a1d8-46f7-8199-6214af4ebc98 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.380654] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 521.380654] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 521.380654] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 521.380654] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 521.380654] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 521.380654] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Didn't find any instances for network info cache update. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 521.385078] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.385078] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.385078] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.385078] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.385078] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.385078] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.385301] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 521.385301] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.408074] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.423701] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "3356d528-14ea-4a7d-acef-6c6f56195f31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.424039] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "3356d528-14ea-4a7d-acef-6c6f56195f31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.448851] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 521.455355] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.601716] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198125, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029947} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.601716] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] File moved {{(pid=62368) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 521.601716] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Cleaning up location [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 521.601716] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleting the datastore file [datastore1] vmware_temp/7d0f97df-1a0c-47f1-bbe8-febd983a0190 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 521.602541] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ffc1b0f-640a-430b-af98-1e539c58be1c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.612043] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 521.612043] env[62368]: value = "task-1198126" [ 521.612043] env[62368]: _type = "Task" [ 521.612043] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.622705] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198126, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.712885] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 521.748859] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 521.749134] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 521.749285] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 521.749940] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 521.749940] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 521.749940] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 521.749940] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 521.751593] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 521.751844] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 521.752032] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 521.752217] env[62368]: DEBUG nova.virt.hardware [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 521.753209] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff0a6d7-4a25-4c05-a6a8-a4dbb62ed37f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.765718] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab89c6bd-ab27-434e-be9c-72b1877d4024 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.887261] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Acquiring lock "73c833ef-7761-4372-9d02-b6dfc945e415" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.892022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Lock "73c833ef-7761-4372-9d02-b6dfc945e415" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.892022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.927269] env[62368]: DEBUG nova.compute.manager [req-2b2af772-8fba-40e5-8b8d-30e90b1fa388 req-b0c44eae-60da-4fdb-a6b8-ce9dc080ee03 service nova] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Received event network-changed-72ed829c-2b51-4d83-87e0-ba3316d3d5ef {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 521.928063] env[62368]: DEBUG nova.compute.manager [req-2b2af772-8fba-40e5-8b8d-30e90b1fa388 req-b0c44eae-60da-4fdb-a6b8-ce9dc080ee03 service nova] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Refreshing instance network info cache due to event network-changed-72ed829c-2b51-4d83-87e0-ba3316d3d5ef. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 521.928260] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b2af772-8fba-40e5-8b8d-30e90b1fa388 req-b0c44eae-60da-4fdb-a6b8-ce9dc080ee03 service nova] Acquiring lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.929364] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 521.958417] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.959067] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 521.959434] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 521.960617] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.963666] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 521.967146] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b2af772-8fba-40e5-8b8d-30e90b1fa388 req-b0c44eae-60da-4fdb-a6b8-ce9dc080ee03 service nova] Acquired lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.967347] env[62368]: DEBUG nova.network.neutron [req-2b2af772-8fba-40e5-8b8d-30e90b1fa388 req-b0c44eae-60da-4fdb-a6b8-ce9dc080ee03 service nova] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Refreshing network info cache for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 521.968057] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b48fe71-b471-45e0-8dd6-6eca7ed80866 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.972573] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.891s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.974070] env[62368]: INFO nova.compute.claims [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 521.994255] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed23a00f-c2e5-42a7-9218-ab575f43977b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.026152] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5c652903-a025-4530-96ee-428da51c8b71 could not be found. [ 522.026386] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 522.026787] env[62368]: INFO nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Took 0.07 seconds to destroy the instance on the hypervisor. [ 522.027046] env[62368]: DEBUG oslo.service.loopingcall [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 522.027707] env[62368]: DEBUG nova.compute.manager [-] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 522.027837] env[62368]: DEBUG nova.network.neutron [-] [instance: 5c652903-a025-4530-96ee-428da51c8b71] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 522.056213] env[62368]: DEBUG nova.network.neutron [-] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 522.124179] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198126, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027563} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.124419] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 522.125155] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12812c76-cdf2-4dc0-ad84-459a95969b28 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.131478] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 522.131478] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52299f23-9f21-f526-9d4f-405f82cb1ae8" [ 522.131478] env[62368]: _type = "Task" [ 522.131478] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.147020] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52299f23-9f21-f526-9d4f-405f82cb1ae8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.393478] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 522.460396] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.482443] env[62368]: DEBUG nova.compute.utils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 522.487518] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 522.487742] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 522.518240] env[62368]: DEBUG nova.network.neutron [req-2b2af772-8fba-40e5-8b8d-30e90b1fa388 req-b0c44eae-60da-4fdb-a6b8-ce9dc080ee03 service nova] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 522.563149] env[62368]: DEBUG nova.network.neutron [-] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.630414] env[62368]: DEBUG nova.policy [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a358bb37b18142df960f768b7fada31b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94c761f204e54ce4805ea598f1f8ec1c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 522.645840] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52299f23-9f21-f526-9d4f-405f82cb1ae8, 'name': SearchDatastore_Task, 'duration_secs': 0.010193} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.648041] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.648041] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 522.648041] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f28be109-c12a-4c91-8365-d5b7611c5ca1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.655426] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 522.655426] env[62368]: value = "task-1198128" [ 522.655426] env[62368]: _type = "Task" [ 522.655426] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.668636] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198128, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.919531] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.925438] env[62368]: DEBUG nova.network.neutron [req-2b2af772-8fba-40e5-8b8d-30e90b1fa388 req-b0c44eae-60da-4fdb-a6b8-ce9dc080ee03 service nova] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.993621] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.065892] env[62368]: INFO nova.compute.manager [-] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Took 1.04 seconds to deallocate network for instance. [ 523.068888] env[62368]: DEBUG nova.compute.claims [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 523.069120] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.135932] env[62368]: ERROR nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. [ 523.135932] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.135932] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.135932] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.135932] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.135932] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.135932] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.135932] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.135932] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.135932] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 523.135932] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.135932] env[62368]: ERROR nova.compute.manager raise self.value [ 523.135932] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.135932] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.135932] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.135932] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.136678] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.136678] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.136678] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. [ 523.136678] env[62368]: ERROR nova.compute.manager [ 523.136678] env[62368]: Traceback (most recent call last): [ 523.136678] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.136678] env[62368]: listener.cb(fileno) [ 523.136678] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.136678] env[62368]: result = function(*args, **kwargs) [ 523.136678] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.136678] env[62368]: return func(*args, **kwargs) [ 523.136678] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.136678] env[62368]: raise e [ 523.136678] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.136678] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 523.136678] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.136678] env[62368]: created_port_ids = self._update_ports_for_instance( [ 523.136678] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.136678] env[62368]: with excutils.save_and_reraise_exception(): [ 523.136678] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.136678] env[62368]: self.force_reraise() [ 523.136678] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.136678] env[62368]: raise self.value [ 523.136678] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.136678] env[62368]: updated_port = self._update_port( [ 523.136678] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.136678] env[62368]: _ensure_no_port_binding_failure(port) [ 523.136678] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.136678] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.137519] env[62368]: nova.exception.PortBindingFailed: Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. [ 523.137519] env[62368]: Removing descriptor: 17 [ 523.137519] env[62368]: ERROR nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Traceback (most recent call last): [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] yield resources [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self.driver.spawn(context, instance, image_meta, [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.137519] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] vm_ref = self.build_virtual_machine(instance, [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] for vif in network_info: [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] return self._sync_wrapper(fn, *args, **kwargs) [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self.wait() [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self[:] = self._gt.wait() [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] return self._exit_event.wait() [ 523.137851] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] result = hub.switch() [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] return self.greenlet.switch() [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] result = function(*args, **kwargs) [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] return func(*args, **kwargs) [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] raise e [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] nwinfo = self.network_api.allocate_for_instance( [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.138208] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] created_port_ids = self._update_ports_for_instance( [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] with excutils.save_and_reraise_exception(): [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self.force_reraise() [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] raise self.value [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] updated_port = self._update_port( [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] _ensure_no_port_binding_failure(port) [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.139199] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] raise exception.PortBindingFailed(port_id=port['id']) [ 523.139595] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] nova.exception.PortBindingFailed: Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. [ 523.139595] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] [ 523.139595] env[62368]: INFO nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Terminating instance [ 523.144261] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.144437] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquired lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.144643] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 523.175166] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198128, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.307471] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f103a46-30cc-48fb-a903-f33039cb04ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.319079] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e1e571-e070-4aff-ba32-1eac98e90dc1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.361851] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7bfacc-99a4-485c-904d-e22932701eb7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.371020] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9639b01f-259a-4f12-a099-900ab046eb7f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.386521] env[62368]: DEBUG nova.compute.provider_tree [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.434442] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b2af772-8fba-40e5-8b8d-30e90b1fa388 req-b0c44eae-60da-4fdb-a6b8-ce9dc080ee03 service nova] Releasing lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.674074] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198128, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724432} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.674074] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 523.674074] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 523.674074] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29ac1712-bf36-4144-9b3b-6c684fc6cf47 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.692020] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 523.692020] env[62368]: value = "task-1198129" [ 523.692020] env[62368]: _type = "Task" [ 523.692020] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.696077] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 523.768765] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "92d0e65f-71c9-4ab8-8994-7425306b92bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.769119] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "92d0e65f-71c9-4ab8-8994-7425306b92bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.890636] env[62368]: DEBUG nova.scheduler.client.report [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.953964] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.004249] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.053497] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.053649] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.053855] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.055379] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.055379] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.055379] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.055379] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.055379] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.058805] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.059137] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.059272] env[62368]: DEBUG nova.virt.hardware [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.060244] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314255ee-8c66-49be-8fec-a6e4f6d1f09a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.071254] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be8419c-3077-4cbf-8831-9ebd7daee7de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.180396] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Successfully created port: 4b89762c-af45-4c1d-aea6-e6058910e094 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.201998] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198129, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132703} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.201998] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 524.201998] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7407ba9a-996b-41c5-be44-965779658130 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.229707] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 524.230314] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ff2b008-8a41-4135-8508-f360a8934bf5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.255252] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 524.255252] env[62368]: value = "task-1198131" [ 524.255252] env[62368]: _type = "Task" [ 524.255252] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.265548] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198131, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.278285] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.395140] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.395681] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 524.399863] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.992s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.402318] env[62368]: INFO nova.compute.claims [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.456995] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Releasing lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.457243] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.457452] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 524.458038] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67bb5207-fdef-4bfa-8e12-aeec804c5d32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.471894] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c8ac51-a7dd-44ba-9636-941a4d224ae2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.498511] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c198d320-9b6a-4177-9e5e-1d577f9a91ce could not be found. [ 524.502021] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 524.502021] env[62368]: INFO nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Took 0.04 seconds to destroy the instance on the hypervisor. [ 524.502021] env[62368]: DEBUG oslo.service.loopingcall [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.502021] env[62368]: DEBUG nova.compute.manager [-] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.502021] env[62368]: DEBUG nova.network.neutron [-] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 524.558987] env[62368]: DEBUG nova.network.neutron [-] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.573296] env[62368]: ERROR nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. [ 524.573296] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.573296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.573296] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.573296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.573296] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.573296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.573296] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.573296] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.573296] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 524.573296] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.573296] env[62368]: ERROR nova.compute.manager raise self.value [ 524.573296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.573296] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.573296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.573296] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.573776] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.573776] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.573776] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. [ 524.573776] env[62368]: ERROR nova.compute.manager [ 524.573776] env[62368]: Traceback (most recent call last): [ 524.573776] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.573776] env[62368]: listener.cb(fileno) [ 524.573776] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.573776] env[62368]: result = function(*args, **kwargs) [ 524.573776] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.573776] env[62368]: return func(*args, **kwargs) [ 524.573776] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.573776] env[62368]: raise e [ 524.573776] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.573776] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 524.573776] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.573776] env[62368]: created_port_ids = self._update_ports_for_instance( [ 524.573776] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.573776] env[62368]: with excutils.save_and_reraise_exception(): [ 524.573776] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.573776] env[62368]: self.force_reraise() [ 524.573776] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.573776] env[62368]: raise self.value [ 524.573776] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.573776] env[62368]: updated_port = self._update_port( [ 524.573776] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.573776] env[62368]: _ensure_no_port_binding_failure(port) [ 524.573776] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.573776] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.574683] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. [ 524.574683] env[62368]: Removing descriptor: 18 [ 524.574683] env[62368]: ERROR nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Traceback (most recent call last): [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] yield resources [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self.driver.spawn(context, instance, image_meta, [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.574683] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] vm_ref = self.build_virtual_machine(instance, [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] for vif in network_info: [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] return self._sync_wrapper(fn, *args, **kwargs) [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self.wait() [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self[:] = self._gt.wait() [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] return self._exit_event.wait() [ 524.575098] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] result = hub.switch() [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] return self.greenlet.switch() [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] result = function(*args, **kwargs) [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] return func(*args, **kwargs) [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] raise e [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] nwinfo = self.network_api.allocate_for_instance( [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.575429] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] created_port_ids = self._update_ports_for_instance( [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] with excutils.save_and_reraise_exception(): [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self.force_reraise() [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] raise self.value [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] updated_port = self._update_port( [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] _ensure_no_port_binding_failure(port) [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.575821] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] raise exception.PortBindingFailed(port_id=port['id']) [ 524.576151] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] nova.exception.PortBindingFailed: Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. [ 524.576151] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] [ 524.576151] env[62368]: INFO nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Terminating instance [ 524.578229] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Acquiring lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.578459] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Acquired lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.578839] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 524.766899] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198131, 'name': ReconfigVM_Task, 'duration_secs': 0.309338} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.767284] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 524.768048] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-093e7407-9be7-43b9-b027-842815b679cd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.775984] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 524.775984] env[62368]: value = "task-1198132" [ 524.775984] env[62368]: _type = "Task" [ 524.775984] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.789843] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198132, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.805503] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.903022] env[62368]: DEBUG nova.compute.utils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 524.904064] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 524.904411] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 525.025975] env[62368]: DEBUG nova.policy [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5708f2a5b624960a96199aab2e7e301', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2bc4ed4028364d8d99017c176ac621d0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.063357] env[62368]: DEBUG nova.network.neutron [-] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.153874] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.289370] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198132, 'name': Rename_Task, 'duration_secs': 0.172157} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.289663] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.295021] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 525.295021] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-371d3f74-6bde-4f99-adac-d528448cb4c2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.304101] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 525.304101] env[62368]: value = "task-1198133" [ 525.304101] env[62368]: _type = "Task" [ 525.304101] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.315215] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198133, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.407972] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 525.569946] env[62368]: INFO nova.compute.manager [-] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Took 1.07 seconds to deallocate network for instance. [ 525.576217] env[62368]: DEBUG nova.compute.claims [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.576393] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.623228] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "54e3c531-9253-4fc0-999c-2cb9a8388d08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.623228] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "54e3c531-9253-4fc0-999c-2cb9a8388d08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.662730] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a5d77a-2200-48ee-adb0-99363c2b7f2c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.673472] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78098af-f083-4082-8cc5-4d8921d6bb21 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.722471] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0770b1-7da3-47cc-b269-194aeba655a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.728717] env[62368]: DEBUG nova.compute.manager [req-5cfe22b1-bbee-4d25-8ced-f7d530ddfd04 req-faf319bd-fe0c-4fdd-a5c7-00e022beb789 service nova] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Received event network-changed-65fbf44f-a1d8-46f7-8199-6214af4ebc98 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.728717] env[62368]: DEBUG nova.compute.manager [req-5cfe22b1-bbee-4d25-8ced-f7d530ddfd04 req-faf319bd-fe0c-4fdd-a5c7-00e022beb789 service nova] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Refreshing instance network info cache due to event network-changed-65fbf44f-a1d8-46f7-8199-6214af4ebc98. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 525.728717] env[62368]: DEBUG oslo_concurrency.lockutils [req-5cfe22b1-bbee-4d25-8ced-f7d530ddfd04 req-faf319bd-fe0c-4fdd-a5c7-00e022beb789 service nova] Acquiring lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.735827] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ebba16d-e8ce-4be9-bf33-45218e401413 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.753330] env[62368]: DEBUG nova.compute.provider_tree [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.756401] env[62368]: DEBUG nova.compute.manager [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Received event network-vif-deleted-72ed829c-2b51-4d83-87e0-ba3316d3d5ef {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.756582] env[62368]: DEBUG nova.compute.manager [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Received event network-changed-f4814bcc-8ada-4a61-82a4-5dd21dd2ca86 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.756738] env[62368]: DEBUG nova.compute.manager [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Refreshing instance network info cache due to event network-changed-f4814bcc-8ada-4a61-82a4-5dd21dd2ca86. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 525.756947] env[62368]: DEBUG oslo_concurrency.lockutils [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] Acquiring lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.757190] env[62368]: DEBUG oslo_concurrency.lockutils [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] Acquired lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.757249] env[62368]: DEBUG nova.network.neutron [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Refreshing network info cache for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 525.795331] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Releasing lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.795608] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.796860] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 525.798267] env[62368]: DEBUG oslo_concurrency.lockutils [req-5cfe22b1-bbee-4d25-8ced-f7d530ddfd04 req-faf319bd-fe0c-4fdd-a5c7-00e022beb789 service nova] Acquired lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.798453] env[62368]: DEBUG nova.network.neutron [req-5cfe22b1-bbee-4d25-8ced-f7d530ddfd04 req-faf319bd-fe0c-4fdd-a5c7-00e022beb789 service nova] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Refreshing network info cache for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 525.799539] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-614b7eaf-0df2-4f78-9535-75cc813a95b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.819715] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198133, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.827338] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee52af8d-68f6-4c13-8cf4-61ff0a90c4c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.866606] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0144a1d0-bc85-4c8c-9342-8c12da4b8967 could not be found. [ 525.866606] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 525.866730] env[62368]: INFO nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Took 0.07 seconds to destroy the instance on the hypervisor. [ 525.866956] env[62368]: DEBUG oslo.service.loopingcall [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.867204] env[62368]: DEBUG nova.compute.manager [-] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.867314] env[62368]: DEBUG nova.network.neutron [-] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 525.938161] env[62368]: DEBUG nova.network.neutron [-] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.163523] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Successfully created port: fa4e60cb-8bd8-498c-b401-8f7ee20a4abd {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.260534] env[62368]: DEBUG nova.scheduler.client.report [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 526.318312] env[62368]: DEBUG nova.network.neutron [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.323296] env[62368]: DEBUG oslo_vmware.api [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198133, 'name': PowerOnVM_Task, 'duration_secs': 0.709302} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.323296] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 526.323443] env[62368]: INFO nova.compute.manager [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Took 9.12 seconds to spawn the instance on the hypervisor. [ 526.323665] env[62368]: DEBUG nova.compute.manager [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 526.324470] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb39e2fc-1662-44e1-8e2b-96bf9124f173 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.384589] env[62368]: DEBUG nova.network.neutron [req-5cfe22b1-bbee-4d25-8ced-f7d530ddfd04 req-faf319bd-fe0c-4fdd-a5c7-00e022beb789 service nova] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.425922] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 526.438619] env[62368]: DEBUG nova.network.neutron [-] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.461886] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 526.461886] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 526.461886] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 526.462100] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 526.465094] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 526.465094] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 526.465094] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 526.465323] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 526.465421] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 526.465657] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 526.465752] env[62368]: DEBUG nova.virt.hardware [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 526.467331] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43724dd-8175-4820-a73c-963473b3ca4b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.477184] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee270f08-66b8-474d-b67b-1f0221733e9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.616218] env[62368]: DEBUG nova.network.neutron [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.656725] env[62368]: DEBUG nova.network.neutron [req-5cfe22b1-bbee-4d25-8ced-f7d530ddfd04 req-faf319bd-fe0c-4fdd-a5c7-00e022beb789 service nova] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.765596] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.766180] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 526.768956] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.877s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.769149] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.769306] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 526.769615] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.310s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.771064] env[62368]: INFO nova.compute.claims [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.774227] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1706df51-a4bc-49ab-bea4-dca5d0d5fb35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.783391] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6799090-de6e-462f-9766-fcc2207e3ad4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.801467] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01f669c-2866-4ca5-afd1-9ddb2be64b46 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.813778] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a6e48a-8dab-4ffb-9d44-6a66a373cb42 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.854925] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181442MB free_disk=157GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 526.854925] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.858790] env[62368]: INFO nova.compute.manager [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Took 15.23 seconds to build instance. [ 526.941399] env[62368]: INFO nova.compute.manager [-] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Took 1.07 seconds to deallocate network for instance. [ 526.944873] env[62368]: DEBUG nova.compute.claims [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.945270] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.119245] env[62368]: DEBUG oslo_concurrency.lockutils [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] Releasing lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.119685] env[62368]: DEBUG nova.compute.manager [req-7a546428-1d57-4afd-acd9-8734af71db80 req-863136ab-add0-4e32-ab1b-aabc4ae6d0fb service nova] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Received event network-vif-deleted-f4814bcc-8ada-4a61-82a4-5dd21dd2ca86 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.161078] env[62368]: DEBUG oslo_concurrency.lockutils [req-5cfe22b1-bbee-4d25-8ced-f7d530ddfd04 req-faf319bd-fe0c-4fdd-a5c7-00e022beb789 service nova] Releasing lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.276076] env[62368]: DEBUG nova.compute.utils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.277527] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 527.277705] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 527.365790] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cda2043e-a3cb-49c0-867b-1f6e31cb732f tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "0eb2a772-7635-481a-ae2c-9f23669dd175" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.746s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.529373] env[62368]: DEBUG nova.policy [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e27b110e0701414298c85adf815fa992', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f8a4f132cf5481298a649eaec5eca73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 527.637436] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Acquiring lock "eac92d1f-d160-4e06-a71c-d064acdffb19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.639894] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Lock "eac92d1f-d160-4e06-a71c-d064acdffb19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.783725] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 527.870666] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.050721] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252cf161-7b4b-4747-9b6d-d4f4c3f99ac9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.064553] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d8e159-94e9-4179-b00c-008804e03389 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.107380] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416766a2-3fd0-481e-be58-8eb509148c22 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.116746] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dcdc545-2199-4469-8885-67ba0047ddc0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.134461] env[62368]: DEBUG nova.compute.provider_tree [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.402543] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.446945] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Successfully created port: 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.643600] env[62368]: DEBUG nova.scheduler.client.report [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 528.691187] env[62368]: ERROR nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 528.691187] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 528.691187] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.691187] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 528.691187] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.691187] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 528.691187] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.691187] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 528.691187] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.691187] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 528.691187] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.691187] env[62368]: ERROR nova.compute.manager raise self.value [ 528.691187] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.691187] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 528.691187] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.691187] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 528.691777] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.691777] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 528.691777] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 528.691777] env[62368]: ERROR nova.compute.manager [ 528.691777] env[62368]: Traceback (most recent call last): [ 528.691777] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 528.691777] env[62368]: listener.cb(fileno) [ 528.691777] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.691777] env[62368]: result = function(*args, **kwargs) [ 528.691777] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.691777] env[62368]: return func(*args, **kwargs) [ 528.691777] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.691777] env[62368]: raise e [ 528.691777] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.691777] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 528.691777] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.691777] env[62368]: created_port_ids = self._update_ports_for_instance( [ 528.691777] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.691777] env[62368]: with excutils.save_and_reraise_exception(): [ 528.691777] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.691777] env[62368]: self.force_reraise() [ 528.691777] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.691777] env[62368]: raise self.value [ 528.691777] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.691777] env[62368]: updated_port = self._update_port( [ 528.691777] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.691777] env[62368]: _ensure_no_port_binding_failure(port) [ 528.691777] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.691777] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 528.692630] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 528.692630] env[62368]: Removing descriptor: 20 [ 528.692630] env[62368]: ERROR nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Traceback (most recent call last): [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] yield resources [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self.driver.spawn(context, instance, image_meta, [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.692630] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] vm_ref = self.build_virtual_machine(instance, [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] for vif in network_info: [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] return self._sync_wrapper(fn, *args, **kwargs) [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self.wait() [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self[:] = self._gt.wait() [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] return self._exit_event.wait() [ 528.693043] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] result = hub.switch() [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] return self.greenlet.switch() [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] result = function(*args, **kwargs) [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] return func(*args, **kwargs) [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] raise e [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] nwinfo = self.network_api.allocate_for_instance( [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.693622] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] created_port_ids = self._update_ports_for_instance( [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] with excutils.save_and_reraise_exception(): [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self.force_reraise() [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] raise self.value [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] updated_port = self._update_port( [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] _ensure_no_port_binding_failure(port) [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.694017] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] raise exception.PortBindingFailed(port_id=port['id']) [ 528.694407] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 528.694407] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] [ 528.694407] env[62368]: INFO nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Terminating instance [ 528.695571] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Acquiring lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.695733] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Acquired lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.695894] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 528.796441] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 528.826733] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 528.827945] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 528.828341] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 528.828614] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 528.828821] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 528.829030] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 528.829310] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 528.829531] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 528.829775] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 528.829992] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 528.830275] env[62368]: DEBUG nova.virt.hardware [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 528.831243] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1c5648-680a-47a4-a764-842a07650f0b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.844945] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d428c4-3e5b-45a2-a104-76d229ccc911 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.054388] env[62368]: DEBUG nova.compute.manager [req-2f55c028-4da3-42ff-9eb0-99fb0f753e1d req-933e223e-bd79-4132-a8e5-920f0d27a58b service nova] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Received event network-vif-deleted-65fbf44f-a1d8-46f7-8199-6214af4ebc98 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.152261] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.152261] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.158306] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.238s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.161313] env[62368]: INFO nova.compute.claims [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.229511] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.256866] env[62368]: DEBUG nova.compute.manager [req-7e7b08d1-4d74-4ac2-8398-a3e01648c5d9 req-57adc9e1-dbc9-4513-981a-0c56483eaec9 service nova] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Received event network-changed-4b89762c-af45-4c1d-aea6-e6058910e094 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.257095] env[62368]: DEBUG nova.compute.manager [req-7e7b08d1-4d74-4ac2-8398-a3e01648c5d9 req-57adc9e1-dbc9-4513-981a-0c56483eaec9 service nova] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Refreshing instance network info cache due to event network-changed-4b89762c-af45-4c1d-aea6-e6058910e094. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 529.257315] env[62368]: DEBUG oslo_concurrency.lockutils [req-7e7b08d1-4d74-4ac2-8398-a3e01648c5d9 req-57adc9e1-dbc9-4513-981a-0c56483eaec9 service nova] Acquiring lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.329665] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.456600] env[62368]: ERROR nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. [ 529.456600] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.456600] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.456600] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.456600] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.456600] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.456600] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.456600] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.456600] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.456600] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 529.456600] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.456600] env[62368]: ERROR nova.compute.manager raise self.value [ 529.456600] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.456600] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.456600] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.456600] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.457090] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.457090] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.457090] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. [ 529.457090] env[62368]: ERROR nova.compute.manager [ 529.457090] env[62368]: Traceback (most recent call last): [ 529.458397] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.458397] env[62368]: listener.cb(fileno) [ 529.458397] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.458397] env[62368]: result = function(*args, **kwargs) [ 529.458397] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.458397] env[62368]: return func(*args, **kwargs) [ 529.458397] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.458397] env[62368]: raise e [ 529.458397] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.458397] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 529.458397] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.458397] env[62368]: created_port_ids = self._update_ports_for_instance( [ 529.458397] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.458397] env[62368]: with excutils.save_and_reraise_exception(): [ 529.458397] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.458397] env[62368]: self.force_reraise() [ 529.458397] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.458397] env[62368]: raise self.value [ 529.458397] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.458397] env[62368]: updated_port = self._update_port( [ 529.458397] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.458397] env[62368]: _ensure_no_port_binding_failure(port) [ 529.458397] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.458397] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.458397] env[62368]: nova.exception.PortBindingFailed: Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. [ 529.458397] env[62368]: Removing descriptor: 18 [ 529.459141] env[62368]: ERROR nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Traceback (most recent call last): [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] yield resources [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self.driver.spawn(context, instance, image_meta, [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] vm_ref = self.build_virtual_machine(instance, [ 529.459141] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] for vif in network_info: [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] return self._sync_wrapper(fn, *args, **kwargs) [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self.wait() [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self[:] = self._gt.wait() [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] return self._exit_event.wait() [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.459695] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] result = hub.switch() [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] return self.greenlet.switch() [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] result = function(*args, **kwargs) [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] return func(*args, **kwargs) [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] raise e [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] nwinfo = self.network_api.allocate_for_instance( [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] created_port_ids = self._update_ports_for_instance( [ 529.460085] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] with excutils.save_and_reraise_exception(): [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self.force_reraise() [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] raise self.value [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] updated_port = self._update_port( [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] _ensure_no_port_binding_failure(port) [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] raise exception.PortBindingFailed(port_id=port['id']) [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] nova.exception.PortBindingFailed: Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. [ 529.460512] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] [ 529.460938] env[62368]: INFO nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Terminating instance [ 529.462902] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Acquiring lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.463240] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Acquired lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.463609] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 529.481645] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Acquiring lock "01654798-4e54-4177-b20c-f6e0aecf72a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.481926] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Lock "01654798-4e54-4177-b20c-f6e0aecf72a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.666121] env[62368]: DEBUG nova.compute.utils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.675832] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.676143] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 529.707828] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "c9fe2763-2d3c-48f6-a836-e0c8634acc95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.707828] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "c9fe2763-2d3c-48f6-a836-e0c8634acc95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.730358] env[62368]: DEBUG nova.policy [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f57279b5eaaa45be83b8026213c91099', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fc4b3b46106480693f6e18f142bb31d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 529.835360] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Releasing lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.835852] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 529.836026] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 529.836350] env[62368]: DEBUG oslo_concurrency.lockutils [req-7e7b08d1-4d74-4ac2-8398-a3e01648c5d9 req-57adc9e1-dbc9-4513-981a-0c56483eaec9 service nova] Acquired lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.836520] env[62368]: DEBUG nova.network.neutron [req-7e7b08d1-4d74-4ac2-8398-a3e01648c5d9 req-57adc9e1-dbc9-4513-981a-0c56483eaec9 service nova] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Refreshing network info cache for port 4b89762c-af45-4c1d-aea6-e6058910e094 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 529.837742] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0191abc-c771-4cee-8563-50f397093eba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.854976] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f952d9f-1f82-41bf-befc-480aebd9708f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.886415] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 50feff45-9ede-4d4d-b136-bfb5fd916dc9 could not be found. [ 529.889026] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 529.889026] env[62368]: INFO nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 529.889026] env[62368]: DEBUG oslo.service.loopingcall [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.889026] env[62368]: DEBUG nova.compute.manager [-] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.889026] env[62368]: DEBUG nova.network.neutron [-] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 529.922601] env[62368]: DEBUG nova.network.neutron [-] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.946640] env[62368]: ERROR nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. [ 529.946640] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.946640] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.946640] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.946640] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.946640] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.946640] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.946640] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.946640] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.946640] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 529.946640] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.946640] env[62368]: ERROR nova.compute.manager raise self.value [ 529.946640] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.946640] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.946640] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.946640] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.947150] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.947150] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.947150] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. [ 529.947150] env[62368]: ERROR nova.compute.manager [ 529.947150] env[62368]: Traceback (most recent call last): [ 529.947150] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.947150] env[62368]: listener.cb(fileno) [ 529.947150] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.947150] env[62368]: result = function(*args, **kwargs) [ 529.947150] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.947150] env[62368]: return func(*args, **kwargs) [ 529.947150] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.947150] env[62368]: raise e [ 529.947150] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.947150] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 529.947150] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.947150] env[62368]: created_port_ids = self._update_ports_for_instance( [ 529.947150] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.947150] env[62368]: with excutils.save_and_reraise_exception(): [ 529.947150] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.947150] env[62368]: self.force_reraise() [ 529.947150] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.947150] env[62368]: raise self.value [ 529.947150] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.947150] env[62368]: updated_port = self._update_port( [ 529.947150] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.947150] env[62368]: _ensure_no_port_binding_failure(port) [ 529.947150] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.947150] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.948031] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. [ 529.948031] env[62368]: Removing descriptor: 15 [ 529.948031] env[62368]: ERROR nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Traceback (most recent call last): [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] yield resources [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self.driver.spawn(context, instance, image_meta, [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.948031] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] vm_ref = self.build_virtual_machine(instance, [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] for vif in network_info: [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] return self._sync_wrapper(fn, *args, **kwargs) [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self.wait() [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self[:] = self._gt.wait() [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] return self._exit_event.wait() [ 529.948406] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] result = hub.switch() [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] return self.greenlet.switch() [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] result = function(*args, **kwargs) [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] return func(*args, **kwargs) [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] raise e [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] nwinfo = self.network_api.allocate_for_instance( [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.949037] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] created_port_ids = self._update_ports_for_instance( [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] with excutils.save_and_reraise_exception(): [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self.force_reraise() [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] raise self.value [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] updated_port = self._update_port( [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] _ensure_no_port_binding_failure(port) [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.949521] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] raise exception.PortBindingFailed(port_id=port['id']) [ 529.949906] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] nova.exception.PortBindingFailed: Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. [ 529.949906] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] [ 529.949906] env[62368]: INFO nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Terminating instance [ 529.951469] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Acquiring lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.951738] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Acquired lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.952124] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 529.986269] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.052157] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.176330] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.366844] env[62368]: DEBUG nova.network.neutron [req-7e7b08d1-4d74-4ac2-8398-a3e01648c5d9 req-57adc9e1-dbc9-4513-981a-0c56483eaec9 service nova] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.406789] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Successfully created port: d03daeea-1059-48e6-bb5e-3e0939733e47 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.425076] env[62368]: DEBUG nova.network.neutron [-] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.440903] env[62368]: DEBUG nova.network.neutron [req-7e7b08d1-4d74-4ac2-8398-a3e01648c5d9 req-57adc9e1-dbc9-4513-981a-0c56483eaec9 service nova] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.445694] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599c3457-8279-4601-a10c-377dc21e4091 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.457023] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45635503-f5db-42ab-8a03-e5e163a0c81f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.493619] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.496532] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8620c704-124c-44d3-9b42-3cd43d423a07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.509993] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf625e6-2c26-4bbb-936f-d177c2428e42 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.527816] env[62368]: DEBUG nova.compute.provider_tree [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.555878] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Releasing lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.556884] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 530.557181] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 530.557550] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b9443d9-7ba9-4246-a68a-5d1b5fe2643c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.569165] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12dbbbc6-1404-46d2-9c6c-be648513ee04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.587389] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Acquiring lock "15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.587684] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Lock "15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.604679] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5d0e30b7-17b1-46c1-9cd0-05df00c9923e could not be found. [ 530.605042] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 530.605288] env[62368]: INFO nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 530.605582] env[62368]: DEBUG oslo.service.loopingcall [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 530.605848] env[62368]: DEBUG nova.compute.manager [-] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 530.605987] env[62368]: DEBUG nova.network.neutron [-] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 530.652074] env[62368]: DEBUG nova.network.neutron [-] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.770621] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.930641] env[62368]: INFO nova.compute.manager [-] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Took 1.04 seconds to deallocate network for instance. [ 530.933660] env[62368]: DEBUG nova.compute.claims [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 530.933873] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.947966] env[62368]: DEBUG oslo_concurrency.lockutils [req-7e7b08d1-4d74-4ac2-8398-a3e01648c5d9 req-57adc9e1-dbc9-4513-981a-0c56483eaec9 service nova] Releasing lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.004629] env[62368]: INFO nova.compute.manager [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Rebuilding instance [ 531.031482] env[62368]: DEBUG nova.scheduler.client.report [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.082986] env[62368]: DEBUG nova.compute.manager [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 531.082986] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40242e1-da43-43c7-8e77-52a7140423b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.154970] env[62368]: DEBUG nova.network.neutron [-] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.195221] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.228312] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.228841] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.229301] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.229543] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.230192] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.230385] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.230659] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.230815] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.231053] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.231277] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.231820] env[62368]: DEBUG nova.virt.hardware [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.232449] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3574f072-d7df-46ac-b4cb-4d92ddf60728 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.245448] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b71e76c-e29a-48e3-b6f7-255314bcfe5b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.276397] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Releasing lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.276907] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 531.277199] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 531.277644] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-631fe771-ccb1-4d93-ac00-d7ad86b4e6d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.289418] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b14cfe-44f6-4b91-887a-47acf9f11fc4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.322154] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4 could not be found. [ 531.322154] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 531.322154] env[62368]: INFO nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 531.322527] env[62368]: DEBUG oslo.service.loopingcall [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 531.322705] env[62368]: DEBUG nova.compute.manager [-] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 531.322946] env[62368]: DEBUG nova.network.neutron [-] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 531.349455] env[62368]: DEBUG nova.network.neutron [-] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.482356] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Acquiring lock "546a40aa-c278-460c-8c91-c6c903d751ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.482356] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Lock "546a40aa-c278-460c-8c91-c6c903d751ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.537414] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.537626] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.540255] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.471s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.596172] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 531.596837] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2806b5b-371a-4815-a7f3-4b3e79c92ffa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.607570] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 531.607570] env[62368]: value = "task-1198135" [ 531.607570] env[62368]: _type = "Task" [ 531.607570] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.618319] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.658553] env[62368]: INFO nova.compute.manager [-] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Took 1.05 seconds to deallocate network for instance. [ 531.661374] env[62368]: DEBUG nova.compute.claims [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 531.661601] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.853788] env[62368]: DEBUG nova.network.neutron [-] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.904172] env[62368]: DEBUG nova.compute.manager [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Received event network-changed-fa4e60cb-8bd8-498c-b401-8f7ee20a4abd {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 531.904172] env[62368]: DEBUG nova.compute.manager [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Refreshing instance network info cache due to event network-changed-fa4e60cb-8bd8-498c-b401-8f7ee20a4abd. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 531.904172] env[62368]: DEBUG oslo_concurrency.lockutils [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] Acquiring lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.904562] env[62368]: DEBUG oslo_concurrency.lockutils [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] Acquired lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.904836] env[62368]: DEBUG nova.network.neutron [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Refreshing network info cache for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 531.938823] env[62368]: DEBUG nova.compute.manager [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Received event network-vif-deleted-4b89762c-af45-4c1d-aea6-e6058910e094 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 531.938952] env[62368]: DEBUG nova.compute.manager [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Received event network-changed-8ec1919a-6e92-4c70-b8b0-9f3b01caeabf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 531.939141] env[62368]: DEBUG nova.compute.manager [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Refreshing instance network info cache due to event network-changed-8ec1919a-6e92-4c70-b8b0-9f3b01caeabf. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 531.939353] env[62368]: DEBUG oslo_concurrency.lockutils [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] Acquiring lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.939695] env[62368]: DEBUG oslo_concurrency.lockutils [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] Acquired lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.939695] env[62368]: DEBUG nova.network.neutron [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Refreshing network info cache for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 532.010142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "fd5478d6-9fa8-4471-8b95-b89f3d826bcc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.010431] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "fd5478d6-9fa8-4471-8b95-b89f3d826bcc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.047226] env[62368]: DEBUG nova.compute.utils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.048751] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.049525] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.121256] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198135, 'name': PowerOffVM_Task, 'duration_secs': 0.223636} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.121256] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 532.121448] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 532.122473] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f7da1f-d2c7-4d01-9014-cc27d83088f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.129686] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 532.133103] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8fa9ea4-05dd-484c-b98f-4cb6d28404e0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.144043] env[62368]: DEBUG nova.policy [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e031564d04948249ad8013de6f63418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1fc8b46f06a0480393978179d80bfa45', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.172224] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 532.172882] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 532.172882] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleting the datastore file [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 532.176717] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0ab707a-f77c-4cf3-8e80-636e7a0ae36c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.187373] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 532.187373] env[62368]: value = "task-1198137" [ 532.187373] env[62368]: _type = "Task" [ 532.187373] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.205846] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.360052] env[62368]: INFO nova.compute.manager [-] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Took 1.04 seconds to deallocate network for instance. [ 532.363258] env[62368]: DEBUG nova.compute.claims [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 532.363480] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.364294] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f43f808-6d7a-4b9d-a480-0b3b1d39d01c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.375334] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd38d0f9-41f7-4abd-a822-f1242788f2b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.417394] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d3071c-5e1e-469f-9aac-db3aa12ec52b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.422624] env[62368]: ERROR nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. [ 532.422624] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.422624] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.422624] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.422624] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.422624] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.422624] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.422624] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.422624] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.422624] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 532.422624] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.422624] env[62368]: ERROR nova.compute.manager raise self.value [ 532.422624] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.422624] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.422624] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.422624] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.423219] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.423219] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.423219] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. [ 532.423219] env[62368]: ERROR nova.compute.manager [ 532.423219] env[62368]: Traceback (most recent call last): [ 532.423219] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.423219] env[62368]: listener.cb(fileno) [ 532.423219] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.423219] env[62368]: result = function(*args, **kwargs) [ 532.423219] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.423219] env[62368]: return func(*args, **kwargs) [ 532.423219] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.423219] env[62368]: raise e [ 532.423219] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.423219] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 532.423219] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.423219] env[62368]: created_port_ids = self._update_ports_for_instance( [ 532.423219] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.423219] env[62368]: with excutils.save_and_reraise_exception(): [ 532.423219] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.423219] env[62368]: self.force_reraise() [ 532.423219] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.423219] env[62368]: raise self.value [ 532.423219] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.423219] env[62368]: updated_port = self._update_port( [ 532.423219] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.423219] env[62368]: _ensure_no_port_binding_failure(port) [ 532.423219] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.423219] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.424106] env[62368]: nova.exception.PortBindingFailed: Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. [ 532.424106] env[62368]: Removing descriptor: 18 [ 532.424106] env[62368]: ERROR nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Traceback (most recent call last): [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] yield resources [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self.driver.spawn(context, instance, image_meta, [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.424106] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] vm_ref = self.build_virtual_machine(instance, [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] for vif in network_info: [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] return self._sync_wrapper(fn, *args, **kwargs) [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self.wait() [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self[:] = self._gt.wait() [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] return self._exit_event.wait() [ 532.424538] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] result = hub.switch() [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] return self.greenlet.switch() [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] result = function(*args, **kwargs) [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] return func(*args, **kwargs) [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] raise e [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] nwinfo = self.network_api.allocate_for_instance( [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.424879] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] created_port_ids = self._update_ports_for_instance( [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] with excutils.save_and_reraise_exception(): [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self.force_reraise() [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] raise self.value [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] updated_port = self._update_port( [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] _ensure_no_port_binding_failure(port) [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.425253] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] raise exception.PortBindingFailed(port_id=port['id']) [ 532.425601] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] nova.exception.PortBindingFailed: Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. [ 532.425601] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] [ 532.425601] env[62368]: INFO nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Terminating instance [ 532.427379] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.427379] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquired lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.427495] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 532.429537] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94181f1-e87a-467b-82da-cf9e31442992 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.447441] env[62368]: DEBUG nova.compute.provider_tree [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 532.450729] env[62368]: DEBUG nova.network.neutron [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.484590] env[62368]: DEBUG nova.network.neutron [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.553025] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.559109] env[62368]: DEBUG nova.network.neutron [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.611340] env[62368]: DEBUG nova.network.neutron [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.702752] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198137, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110077} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.703026] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 532.703209] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 532.703381] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 532.760406] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Successfully created port: 4e18f0de-825f-466a-80b7-7bb7a8333db5 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.954227] env[62368]: DEBUG nova.scheduler.client.report [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 532.965995] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.065237] env[62368]: DEBUG oslo_concurrency.lockutils [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] Releasing lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.065480] env[62368]: DEBUG nova.compute.manager [req-193819d1-f749-4149-a104-75cb9f341e6a req-9070c7f0-7f60-4421-8a3b-a66d809192f4 service nova] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Received event network-vif-deleted-fa4e60cb-8bd8-498c-b401-8f7ee20a4abd {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.117063] env[62368]: DEBUG oslo_concurrency.lockutils [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] Releasing lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.117324] env[62368]: DEBUG nova.compute.manager [req-21ad9008-ec2b-4f2a-8f73-4b6dfb8bea8d req-caf00cfa-a5b0-436e-b70f-212fc9d45c72 service nova] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Received event network-vif-deleted-8ec1919a-6e92-4c70-b8b0-9f3b01caeabf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.156128] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.187571] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "fc4267ff-dda1-444f-a5b0-a4954626b42d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.187859] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "fc4267ff-dda1-444f-a5b0-a4954626b42d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.461652] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.462333] env[62368]: ERROR nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] Traceback (most recent call last): [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self.driver.spawn(context, instance, image_meta, [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] vm_ref = self.build_virtual_machine(instance, [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.462333] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] for vif in network_info: [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] return self._sync_wrapper(fn, *args, **kwargs) [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self.wait() [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self[:] = self._gt.wait() [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] return self._exit_event.wait() [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] result = hub.switch() [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.462682] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] return self.greenlet.switch() [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] result = function(*args, **kwargs) [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] return func(*args, **kwargs) [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] raise e [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] nwinfo = self.network_api.allocate_for_instance( [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] created_port_ids = self._update_ports_for_instance( [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] with excutils.save_and_reraise_exception(): [ 533.463048] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] self.force_reraise() [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] raise self.value [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] updated_port = self._update_port( [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] _ensure_no_port_binding_failure(port) [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] raise exception.PortBindingFailed(port_id=port['id']) [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] nova.exception.PortBindingFailed: Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. [ 533.463401] env[62368]: ERROR nova.compute.manager [instance: 5c652903-a025-4530-96ee-428da51c8b71] [ 533.463709] env[62368]: DEBUG nova.compute.utils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 533.465113] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.659s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.467520] env[62368]: INFO nova.compute.claims [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 533.480334] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Build of instance 5c652903-a025-4530-96ee-428da51c8b71 was re-scheduled: Binding failed for port 72ed829c-2b51-4d83-87e0-ba3316d3d5ef, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 533.482044] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 533.482044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.482044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.482044] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.567117] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.615663] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.617460] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.617460] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.617460] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.617460] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.617460] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.617781] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.617781] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.617781] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.622754] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.622754] env[62368]: DEBUG nova.virt.hardware [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.623110] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7995c1-6e34-458f-b953-1d11297475ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.635127] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fc05a3-12bf-43f6-83ed-05dd560a0568 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.663216] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Releasing lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.663216] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 533.663216] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 533.663216] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24ddc452-3a60-49cf-8973-43d4f9c25c9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.673159] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bc2d49-f380-4be1-b614-7f40497d3184 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.697810] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3356d528-14ea-4a7d-acef-6c6f56195f31 could not be found. [ 533.698029] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 533.699016] env[62368]: INFO nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Took 0.04 seconds to destroy the instance on the hypervisor. [ 533.699016] env[62368]: DEBUG oslo.service.loopingcall [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.699016] env[62368]: DEBUG nova.compute.manager [-] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.699016] env[62368]: DEBUG nova.network.neutron [-] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 533.721441] env[62368]: DEBUG nova.network.neutron [-] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.745047] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.745047] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.745229] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.745329] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.745463] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.745608] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.745812] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.745969] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.746869] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.746869] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.746869] env[62368]: DEBUG nova.virt.hardware [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.747715] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503edaa2-c269-4eb7-983e-ea0a81ce29b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.756742] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0c5a23-9959-426a-81cb-34702e95c6fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.771290] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 533.776834] env[62368]: DEBUG oslo.service.loopingcall [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.777455] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 533.777674] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ed44d00-fea3-4126-a22d-70130b98d8f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.795797] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 533.795797] env[62368]: value = "task-1198138" [ 533.795797] env[62368]: _type = "Task" [ 533.795797] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.804396] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198138, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.003157] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.102354] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.223937] env[62368]: DEBUG nova.network.neutron [-] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.307870] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198138, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.507748] env[62368]: DEBUG nova.compute.manager [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Received event network-changed-d03daeea-1059-48e6-bb5e-3e0939733e47 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.508483] env[62368]: DEBUG nova.compute.manager [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Refreshing instance network info cache due to event network-changed-d03daeea-1059-48e6-bb5e-3e0939733e47. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 534.508483] env[62368]: DEBUG oslo_concurrency.lockutils [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] Acquiring lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.508483] env[62368]: DEBUG oslo_concurrency.lockutils [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] Acquired lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.508711] env[62368]: DEBUG nova.network.neutron [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Refreshing network info cache for port d03daeea-1059-48e6-bb5e-3e0939733e47 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 534.605442] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "refresh_cache-5c652903-a025-4530-96ee-428da51c8b71" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.605442] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 534.605617] env[62368]: DEBUG nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.605665] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 534.657505] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.730831] env[62368]: INFO nova.compute.manager [-] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Took 1.03 seconds to deallocate network for instance. [ 534.739409] env[62368]: DEBUG nova.compute.claims [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 534.739537] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.771836] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f6aade-578c-4514-b998-ec44bf29f691 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.782089] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f834993-21c7-420b-8dd7-383212ff5dfa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.824552] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec7207f-fa33-499b-b603-4c9d12952c26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.837733] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198138, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.840390] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f125fc0b-433c-4ae9-80cf-55b84e417ca7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.860824] env[62368]: DEBUG nova.compute.provider_tree [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.059112] env[62368]: DEBUG nova.network.neutron [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.116809] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Acquiring lock "489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.117050] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Lock "489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.161503] env[62368]: DEBUG nova.network.neutron [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.329826] env[62368]: DEBUG nova.network.neutron [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.340438] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198138, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.364142] env[62368]: DEBUG nova.scheduler.client.report [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.653254] env[62368]: DEBUG nova.compute.manager [req-2eaac728-7e08-4e42-a7e3-c07c5f8acbef req-985ecac7-32cb-4714-bb80-67a35c996622 service nova] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Received event network-changed-4e18f0de-825f-466a-80b7-7bb7a8333db5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.653501] env[62368]: DEBUG nova.compute.manager [req-2eaac728-7e08-4e42-a7e3-c07c5f8acbef req-985ecac7-32cb-4714-bb80-67a35c996622 service nova] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Refreshing instance network info cache due to event network-changed-4e18f0de-825f-466a-80b7-7bb7a8333db5. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 535.653882] env[62368]: DEBUG oslo_concurrency.lockutils [req-2eaac728-7e08-4e42-a7e3-c07c5f8acbef req-985ecac7-32cb-4714-bb80-67a35c996622 service nova] Acquiring lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.653882] env[62368]: DEBUG oslo_concurrency.lockutils [req-2eaac728-7e08-4e42-a7e3-c07c5f8acbef req-985ecac7-32cb-4714-bb80-67a35c996622 service nova] Acquired lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.654325] env[62368]: DEBUG nova.network.neutron [req-2eaac728-7e08-4e42-a7e3-c07c5f8acbef req-985ecac7-32cb-4714-bb80-67a35c996622 service nova] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Refreshing network info cache for port 4e18f0de-825f-466a-80b7-7bb7a8333db5 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 535.668614] env[62368]: INFO nova.compute.manager [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 5c652903-a025-4530-96ee-428da51c8b71] Took 1.06 seconds to deallocate network for instance. [ 535.723419] env[62368]: ERROR nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. [ 535.723419] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 535.723419] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.723419] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 535.723419] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.723419] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 535.723419] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.723419] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 535.723419] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.723419] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 535.723419] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.723419] env[62368]: ERROR nova.compute.manager raise self.value [ 535.723419] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.723419] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 535.723419] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.723419] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 535.723906] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.723906] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 535.723906] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. [ 535.723906] env[62368]: ERROR nova.compute.manager [ 535.723906] env[62368]: Traceback (most recent call last): [ 535.723906] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 535.723906] env[62368]: listener.cb(fileno) [ 535.723906] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.723906] env[62368]: result = function(*args, **kwargs) [ 535.723906] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.723906] env[62368]: return func(*args, **kwargs) [ 535.723906] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.723906] env[62368]: raise e [ 535.723906] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.723906] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 535.723906] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.723906] env[62368]: created_port_ids = self._update_ports_for_instance( [ 535.723906] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.723906] env[62368]: with excutils.save_and_reraise_exception(): [ 535.723906] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.723906] env[62368]: self.force_reraise() [ 535.723906] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.723906] env[62368]: raise self.value [ 535.723906] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.723906] env[62368]: updated_port = self._update_port( [ 535.723906] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.723906] env[62368]: _ensure_no_port_binding_failure(port) [ 535.723906] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.723906] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 535.725577] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. [ 535.725577] env[62368]: Removing descriptor: 15 [ 535.725577] env[62368]: ERROR nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Traceback (most recent call last): [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] yield resources [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self.driver.spawn(context, instance, image_meta, [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.725577] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] vm_ref = self.build_virtual_machine(instance, [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] for vif in network_info: [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] return self._sync_wrapper(fn, *args, **kwargs) [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self.wait() [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self[:] = self._gt.wait() [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] return self._exit_event.wait() [ 535.726245] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] result = hub.switch() [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] return self.greenlet.switch() [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] result = function(*args, **kwargs) [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] return func(*args, **kwargs) [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] raise e [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] nwinfo = self.network_api.allocate_for_instance( [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.726626] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] created_port_ids = self._update_ports_for_instance( [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] with excutils.save_and_reraise_exception(): [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self.force_reraise() [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] raise self.value [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] updated_port = self._update_port( [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] _ensure_no_port_binding_failure(port) [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.727086] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] raise exception.PortBindingFailed(port_id=port['id']) [ 535.728091] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] nova.exception.PortBindingFailed: Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. [ 535.728091] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] [ 535.728091] env[62368]: INFO nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Terminating instance [ 535.728480] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Acquiring lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.834219] env[62368]: DEBUG oslo_concurrency.lockutils [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] Releasing lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.834478] env[62368]: DEBUG nova.compute.manager [req-8c8356ea-74b9-40cd-b6eb-bf8ac2076534 req-16b75f94-eb21-4efb-8c60-ecfce588148c service nova] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Received event network-vif-deleted-d03daeea-1059-48e6-bb5e-3e0939733e47 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.834807] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198138, 'name': CreateVM_Task, 'duration_secs': 1.64933} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.834897] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 535.836438] env[62368]: DEBUG oslo_vmware.service [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a957d9aa-4638-47df-8f22-8201a418fb7e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.842462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.842885] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.842975] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 535.843281] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56f4f576-ba5f-4499-ac5d-0579c50ad561 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.848344] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 535.848344] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ee518d-8967-7d90-8db0-aaf9df05e842" [ 535.848344] env[62368]: _type = "Task" [ 535.848344] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.858827] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ee518d-8967-7d90-8db0-aaf9df05e842, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.871754] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.872349] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 535.875318] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.299s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.181821] env[62368]: DEBUG nova.network.neutron [req-2eaac728-7e08-4e42-a7e3-c07c5f8acbef req-985ecac7-32cb-4714-bb80-67a35c996622 service nova] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.250279] env[62368]: DEBUG nova.network.neutron [req-2eaac728-7e08-4e42-a7e3-c07c5f8acbef req-985ecac7-32cb-4714-bb80-67a35c996622 service nova] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.361694] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.362029] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 536.362317] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.362494] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.362720] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 536.363378] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56559a31-f8e3-4532-9d56-6c93a4a5f50c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.381268] env[62368]: DEBUG nova.compute.utils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 536.391299] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 536.394124] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 536.400994] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 536.401047] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 536.402817] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba5df07-845b-40da-9416-39b6c50be9a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.414503] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ce97148-a930-4b1c-8cd6-d01b05952311 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.424389] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 536.424389] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52b24e74-e075-7d99-3007-1b3d8d43e9de" [ 536.424389] env[62368]: _type = "Task" [ 536.424389] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.443067] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Preparing fetch location {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 536.443446] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating directory with path [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 536.444042] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2382d15b-5f05-48ab-99ca-2354e2abb3cd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.469599] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Created directory with path [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 536.470151] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Fetch image to [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 536.470236] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Downloading image file data d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d to [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk on the data store datastore2 {{(pid=62368) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 536.471917] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb932a00-8b21-4462-bc0b-1ecd3ab8da9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.480995] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e0f6e0-abad-48e3-bb67-aa6c45ab8a10 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.499205] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff11acba-7dce-4e67-8d37-50d719401a74 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.537778] env[62368]: DEBUG nova.policy [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f57279b5eaaa45be83b8026213c91099', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fc4b3b46106480693f6e18f142bb31d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 536.543027] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2ef9c6-7d44-48f0-9b51-b8a4421b706b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.550742] env[62368]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-b7597415-b35f-4df6-8457-1ecc5a8ad61d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.577857] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Downloading image file data d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d to the data store datastore2 {{(pid=62368) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 536.669202] env[62368]: DEBUG oslo_vmware.rw_handles [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62368) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 536.752697] env[62368]: DEBUG oslo_concurrency.lockutils [req-2eaac728-7e08-4e42-a7e3-c07c5f8acbef req-985ecac7-32cb-4714-bb80-67a35c996622 service nova] Releasing lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.753230] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Acquired lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.753360] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.800174] env[62368]: INFO nova.scheduler.client.report [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted allocations for instance 5c652903-a025-4530-96ee-428da51c8b71 [ 536.889743] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 536.961034] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ad2aed-bbdf-4cf9-aa12-3d082103461e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.974088] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165ae0fd-e9a9-443f-b8e1-817703b1f9e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.015375] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb75352-9565-44ed-8257-9e960d7ecebb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.025384] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Acquiring lock "1523daf8-033b-48e2-a495-885feb4e0ea3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.026766] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Lock "1523daf8-033b-48e2-a495-885feb4e0ea3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.033642] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12083a8a-ba0c-44f7-9cb5-c3ecef8fecf0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.056136] env[62368]: DEBUG nova.compute.provider_tree [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.058165] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Successfully created port: 5c4de9fa-fd2c-4559-b912-cfc173677a4b {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.312225] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3f474599-99a8-4314-a08e-09611b01b0d3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "5c652903-a025-4530-96ee-428da51c8b71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.699s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.313544] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "5c652903-a025-4530-96ee-428da51c8b71" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 24.262s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.314070] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df74ae48-6f0d-4b45-9463-44cb2f59c96a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.330343] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e336689a-beb5-480b-895e-4b19dfb4879d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.422835] env[62368]: DEBUG oslo_vmware.rw_handles [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Completed reading data from the image iterator. {{(pid=62368) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 537.423143] env[62368]: DEBUG oslo_vmware.rw_handles [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 537.459672] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.548813] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Downloaded image file data d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d to vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk on the data store datastore2 {{(pid=62368) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 537.551519] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Caching image {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 537.551519] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Copying Virtual Disk [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk to [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 537.551519] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-19015f3b-79fa-4677-bda8-67b35d756693 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.560941] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 537.560941] env[62368]: value = "task-1198139" [ 537.560941] env[62368]: _type = "Task" [ 537.560941] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.564666] env[62368]: DEBUG nova.scheduler.client.report [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.575669] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198139, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.667064] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.815216] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 537.867117] env[62368]: INFO nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 5c652903-a025-4530-96ee-428da51c8b71] During the sync_power process the instance has moved from host None to host cpu-1 [ 537.867407] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "5c652903-a025-4530-96ee-428da51c8b71" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.554s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.905424] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 537.965863] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.966294] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.966590] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.966590] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.966904] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.966904] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.971717] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.971717] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.971949] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.972745] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.975115] env[62368]: DEBUG nova.virt.hardware [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.979022] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71639b6-b1d5-4d58-bfe4-b2940046b07c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.990646] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0856bf36-c221-4369-83be-9d2809280e27 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.074162] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.198s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.074162] env[62368]: ERROR nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. [ 538.074162] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Traceback (most recent call last): [ 538.074162] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.074162] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self.driver.spawn(context, instance, image_meta, [ 538.074162] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 538.074162] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.074162] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.074162] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] vm_ref = self.build_virtual_machine(instance, [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] for vif in network_info: [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] return self._sync_wrapper(fn, *args, **kwargs) [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self.wait() [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self[:] = self._gt.wait() [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] return self._exit_event.wait() [ 538.074718] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] result = hub.switch() [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] return self.greenlet.switch() [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] result = function(*args, **kwargs) [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] return func(*args, **kwargs) [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] raise e [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] nwinfo = self.network_api.allocate_for_instance( [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.075338] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] created_port_ids = self._update_ports_for_instance( [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] with excutils.save_and_reraise_exception(): [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] self.force_reraise() [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] raise self.value [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] updated_port = self._update_port( [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] _ensure_no_port_binding_failure(port) [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.075959] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] raise exception.PortBindingFailed(port_id=port['id']) [ 538.076493] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] nova.exception.PortBindingFailed: Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. [ 538.076493] env[62368]: ERROR nova.compute.manager [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] [ 538.076493] env[62368]: DEBUG nova.compute.utils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 538.080048] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Build of instance c198d320-9b6a-4177-9e5e-1d577f9a91ce was re-scheduled: Binding failed for port f4814bcc-8ada-4a61-82a4-5dd21dd2ca86, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 538.080048] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 538.080048] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.080048] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquired lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.080551] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 538.080551] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198139, 'name': CopyVirtualDisk_Task} progress is 27%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.080551] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.225s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.169674] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Releasing lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.170288] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.170387] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 538.170666] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fc9b003-e62b-45a3-ba61-50b1d94f2df3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.187718] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94df173-1943-4155-b1d5-5288f33650db {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.224184] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 73c833ef-7761-4372-9d02-b6dfc945e415 could not be found. [ 538.224292] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 538.224470] env[62368]: INFO nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Took 0.05 seconds to destroy the instance on the hypervisor. [ 538.224767] env[62368]: DEBUG oslo.service.loopingcall [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.225170] env[62368]: DEBUG nova.compute.manager [-] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.225292] env[62368]: DEBUG nova.network.neutron [-] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 538.275792] env[62368]: DEBUG nova.network.neutron [-] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.309923] env[62368]: DEBUG nova.compute.manager [req-85ad813a-7389-43f0-a2eb-22798ac26b69 req-d104726c-b74d-4855-946d-7c8ef062f4ed service nova] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Received event network-vif-deleted-4e18f0de-825f-466a-80b7-7bb7a8333db5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.353672] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.578241] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198139, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.651669] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.780268] env[62368]: DEBUG nova.network.neutron [-] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.850301] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.077913] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198139, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.079955} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.078591] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Copied Virtual Disk [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk to [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 539.078591] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleting the datastore file [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/tmp-sparse.vmdk {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 539.078906] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74877e11-2d34-432e-ad0c-cb64b64b8b4b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.088349] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 539.088349] env[62368]: value = "task-1198140" [ 539.088349] env[62368]: _type = "Task" [ 539.088349] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.109127] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.130962] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 0eb2a772-7635-481a-ae2c-9f23669dd175 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 539.289646] env[62368]: INFO nova.compute.manager [-] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Took 1.06 seconds to deallocate network for instance. [ 539.298152] env[62368]: DEBUG nova.compute.claims [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.298152] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.353471] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Releasing lock "refresh_cache-c198d320-9b6a-4177-9e5e-1d577f9a91ce" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.353471] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 539.355673] env[62368]: DEBUG nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.356279] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 539.379315] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.582192] env[62368]: ERROR nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. [ 539.582192] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 539.582192] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.582192] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 539.582192] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.582192] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 539.582192] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.582192] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 539.582192] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.582192] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 539.582192] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.582192] env[62368]: ERROR nova.compute.manager raise self.value [ 539.582192] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.582192] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 539.582192] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.582192] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 539.582842] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.582842] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 539.582842] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. [ 539.582842] env[62368]: ERROR nova.compute.manager [ 539.584346] env[62368]: Traceback (most recent call last): [ 539.584346] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 539.584346] env[62368]: listener.cb(fileno) [ 539.584346] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.584346] env[62368]: result = function(*args, **kwargs) [ 539.584346] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.584346] env[62368]: return func(*args, **kwargs) [ 539.584346] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.584346] env[62368]: raise e [ 539.584346] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.584346] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 539.584346] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.584346] env[62368]: created_port_ids = self._update_ports_for_instance( [ 539.584346] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.584346] env[62368]: with excutils.save_and_reraise_exception(): [ 539.584346] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.584346] env[62368]: self.force_reraise() [ 539.584346] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.584346] env[62368]: raise self.value [ 539.584346] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.584346] env[62368]: updated_port = self._update_port( [ 539.584346] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.584346] env[62368]: _ensure_no_port_binding_failure(port) [ 539.584346] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.584346] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 539.584346] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. [ 539.584346] env[62368]: Removing descriptor: 15 [ 539.585221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquiring lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.585393] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.586466] env[62368]: ERROR nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Traceback (most recent call last): [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] yield resources [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self.driver.spawn(context, instance, image_meta, [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] vm_ref = self.build_virtual_machine(instance, [ 539.586466] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] for vif in network_info: [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] return self._sync_wrapper(fn, *args, **kwargs) [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self.wait() [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self[:] = self._gt.wait() [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] return self._exit_event.wait() [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.587293] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] result = hub.switch() [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] return self.greenlet.switch() [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] result = function(*args, **kwargs) [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] return func(*args, **kwargs) [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] raise e [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] nwinfo = self.network_api.allocate_for_instance( [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] created_port_ids = self._update_ports_for_instance( [ 539.587852] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] with excutils.save_and_reraise_exception(): [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self.force_reraise() [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] raise self.value [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] updated_port = self._update_port( [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] _ensure_no_port_binding_failure(port) [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] raise exception.PortBindingFailed(port_id=port['id']) [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] nova.exception.PortBindingFailed: Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. [ 539.589427] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] [ 539.589928] env[62368]: INFO nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Terminating instance [ 539.589928] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.590337] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquired lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.590337] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 539.605830] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.055516} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.609365] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 539.610734] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Moving file from [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d to [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d. {{(pid=62368) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 539.612560] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-7b455a95-63b2-4487-91ee-798908268e32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.623458] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 539.623458] env[62368]: value = "task-1198141" [ 539.623458] env[62368]: _type = "Task" [ 539.623458] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.635556] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance c198d320-9b6a-4177-9e5e-1d577f9a91ce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 539.636851] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 0144a1d0-bc85-4c8c-9342-8c12da4b8967 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 539.636851] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 50feff45-9ede-4d4d-b136-bfb5fd916dc9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 539.636851] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 5d0e30b7-17b1-46c1-9cd0-05df00c9923e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 539.636851] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 539.637166] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 3356d528-14ea-4a7d-acef-6c6f56195f31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 539.637166] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 73c833ef-7761-4372-9d02-b6dfc945e415 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 539.637166] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 92d0e65f-71c9-4ab8-8994-7425306b92bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 539.642152] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198141, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.883132] env[62368]: DEBUG nova.network.neutron [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.117323] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.136846] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198141, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.030853} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.137242] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] File moved {{(pid=62368) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 540.137519] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Cleaning up location [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 540.137740] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleting the datastore file [datastore2] vmware_temp/b4fdb3b5-3d63-4d87-bc8f-391bd60a202e {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 540.138205] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9528e22-3fb1-42f2-8be1-96a9e0fb91e0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.143840] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 54e3c531-9253-4fc0-999c-2cb9a8388d08 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 540.147284] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 540.147284] env[62368]: value = "task-1198142" [ 540.147284] env[62368]: _type = "Task" [ 540.147284] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.161182] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198142, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.316319] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.385987] env[62368]: INFO nova.compute.manager [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: c198d320-9b6a-4177-9e5e-1d577f9a91ce] Took 1.03 seconds to deallocate network for instance. [ 540.650522] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance eac92d1f-d160-4e06-a71c-d064acdffb19 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 540.661042] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198142, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027961} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.662101] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 540.662272] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfb6b64a-3aad-4061-86aa-702183cbf949 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.669327] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 540.669327] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522cf41d-ca09-2e88-8918-85628bd61ca7" [ 540.669327] env[62368]: _type = "Task" [ 540.669327] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.682463] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522cf41d-ca09-2e88-8918-85628bd61ca7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.821775] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Releasing lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.822489] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 540.823106] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 540.823506] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77b4157d-1b22-4f38-b0b0-b84c4cd1662e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.833366] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "8c179402-a881-4b89-89d7-5e9031ba78cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.833769] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "8c179402-a881-4b89-89d7-5e9031ba78cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.842147] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e67505-705d-4d32-8018-e427d233be9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.873989] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92d0e65f-71c9-4ab8-8994-7425306b92bb could not be found. [ 540.873989] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 540.874223] env[62368]: INFO nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 540.875602] env[62368]: DEBUG oslo.service.loopingcall [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.875602] env[62368]: DEBUG nova.compute.manager [-] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.875602] env[62368]: DEBUG nova.network.neutron [-] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 540.925063] env[62368]: DEBUG nova.network.neutron [-] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.076350] env[62368]: DEBUG nova.compute.manager [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Received event network-changed-5c4de9fa-fd2c-4559-b912-cfc173677a4b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.076350] env[62368]: DEBUG nova.compute.manager [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Refreshing instance network info cache due to event network-changed-5c4de9fa-fd2c-4559-b912-cfc173677a4b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 541.076350] env[62368]: DEBUG oslo_concurrency.lockutils [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] Acquiring lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.076350] env[62368]: DEBUG oslo_concurrency.lockutils [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] Acquired lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.076350] env[62368]: DEBUG nova.network.neutron [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Refreshing network info cache for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 541.157255] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 01654798-4e54-4177-b20c-f6e0aecf72a7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 541.185060] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522cf41d-ca09-2e88-8918-85628bd61ca7, 'name': SearchDatastore_Task, 'duration_secs': 0.00966} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.185060] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.185212] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 541.185408] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93471582-3ccd-4ddb-8608-98a9d84a1dfd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.195117] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 541.195117] env[62368]: value = "task-1198143" [ 541.195117] env[62368]: _type = "Task" [ 541.195117] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.205914] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.428095] env[62368]: DEBUG nova.network.neutron [-] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.442621] env[62368]: INFO nova.scheduler.client.report [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Deleted allocations for instance c198d320-9b6a-4177-9e5e-1d577f9a91ce [ 541.605171] env[62368]: DEBUG nova.network.neutron [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.662304] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance c9fe2763-2d3c-48f6-a836-e0c8634acc95 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 541.710108] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198143, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460279} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.710376] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 541.711027] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 541.711027] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93799de7-94ed-4b5d-9b9f-a1801bbf6a15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.718594] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 541.718594] env[62368]: value = "task-1198144" [ 541.718594] env[62368]: _type = "Task" [ 541.718594] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.728447] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.819262] env[62368]: DEBUG nova.network.neutron [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.931210] env[62368]: INFO nova.compute.manager [-] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Took 1.06 seconds to deallocate network for instance. [ 541.934958] env[62368]: DEBUG nova.compute.claims [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.937140] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.961221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-948dacf8-9a5f-4f1d-8fc9-3a1edaa1c955 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "c198d320-9b6a-4177-9e5e-1d577f9a91ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.020s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.170353] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 542.230692] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066818} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.231062] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 542.231898] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09496a4e-812f-4342-9b1e-2fc948c64647 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.255948] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 542.257779] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06fe172a-7196-447c-aa66-107fa0b618e3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.279360] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 542.279360] env[62368]: value = "task-1198145" [ 542.279360] env[62368]: _type = "Task" [ 542.279360] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.290951] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198145, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.320912] env[62368]: DEBUG oslo_concurrency.lockutils [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] Releasing lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.321235] env[62368]: DEBUG nova.compute.manager [req-e4508b76-4d04-4742-ad21-8d74f2fac2f7 req-87c4a689-8a7e-41e6-a6c4-37ea77ab89b3 service nova] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Received event network-vif-deleted-5c4de9fa-fd2c-4559-b912-cfc173677a4b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.466667] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.673263] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 546a40aa-c278-460c-8c91-c6c903d751ab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 542.794038] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198145, 'name': ReconfigVM_Task, 'duration_secs': 0.298512} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.794420] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 542.795366] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2b173ad-ea12-4b29-bc51-5c32ddd9e618 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.804605] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 542.804605] env[62368]: value = "task-1198146" [ 542.804605] env[62368]: _type = "Task" [ 542.804605] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.815878] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198146, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.001381] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.069034] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "45fe27f5-0f20-48ff-ad25-03862ac9bdd9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.069311] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "45fe27f5-0f20-48ff-ad25-03862ac9bdd9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.179221] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance fd5478d6-9fa8-4471-8b95-b89f3d826bcc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 543.319194] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198146, 'name': Rename_Task, 'duration_secs': 0.165084} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.319481] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 543.319812] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a5c6a9d-1c32-4f4c-8ed8-dd2b5420ef62 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.327780] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 543.327780] env[62368]: value = "task-1198147" [ 543.327780] env[62368]: _type = "Task" [ 543.327780] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.336775] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198147, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.681586] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance fc4267ff-dda1-444f-a5b0-a4954626b42d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 543.839720] env[62368]: DEBUG oslo_vmware.api [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198147, 'name': PowerOnVM_Task, 'duration_secs': 0.484943} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.839974] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 543.840182] env[62368]: DEBUG nova.compute.manager [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 543.840954] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88713014-92ad-4d89-a53e-b5b818669454 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.186028] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 544.359616] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.634484] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "bebed5b9-d067-4cf2-a203-48763685ecb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.634484] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "bebed5b9-d067-4cf2-a203-48763685ecb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.691644] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 1523daf8-033b-48e2-a495-885feb4e0ea3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 544.691644] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 544.691644] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 545.109583] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9ac65f-94bc-4370-8795-7e166189e363 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.118770] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d10893b-a079-454c-8f5d-b45f095097d4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.155317] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7286aa-3c92-49ae-a9f0-14d6850ba616 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.167159] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2395818-606d-4c2f-becd-69ef61071389 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.181444] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.684183] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.721458] env[62368]: INFO nova.compute.manager [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Rebuilding instance [ 545.782195] env[62368]: DEBUG nova.compute.manager [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 545.783173] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8faafd-6d86-4ea0-b6e7-bbdf5cd2f438 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.121782] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Acquiring lock "63c88ac7-9700-4476-9039-feb034de659d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.122030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Lock "63c88ac7-9700-4476-9039-feb034de659d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.190097] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 546.190097] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.109s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.190097] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.244s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.295380] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 546.295380] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68e6d6f9-1066-4523-80b5-bdaad23c115c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.304132] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 546.304132] env[62368]: value = "task-1198148" [ 546.304132] env[62368]: _type = "Task" [ 546.304132] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.314996] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.825589] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198148, 'name': PowerOffVM_Task, 'duration_secs': 0.130619} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.828746] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 546.828979] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 546.834087] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f13b507-7968-48c1-91c1-26f0ce869ae9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.842865] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 546.842865] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25f60fdf-f241-46d2-9024-061e0c68fd32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.875208] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 546.875208] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 546.875208] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Deleting the datastore file [datastore2] 0eb2a772-7635-481a-ae2c-9f23669dd175 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 546.875208] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d9c791d-6b0c-47a6-9ead-8ea5ebe0b574 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.887321] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 546.887321] env[62368]: value = "task-1198150" [ 546.887321] env[62368]: _type = "Task" [ 546.887321] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.906839] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.133446] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "b0bcd561-35cd-475f-975e-e78456a87fac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.133446] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "b0bcd561-35cd-475f-975e-e78456a87fac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.150053] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4b05c7-2e7a-48b8-b26d-6143aaa55eb9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.159809] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5a9dd7-db40-405a-961c-457dda94ce73 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.194758] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a3ae20-809c-4d0b-b936-2c304d082019 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.204740] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af071b6-9f19-4511-9c2e-67851964c5ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.222584] env[62368]: DEBUG nova.compute.provider_tree [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.402235] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126328} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.404135] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 547.404262] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 547.404392] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 547.732191] env[62368]: DEBUG nova.scheduler.client.report [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.240554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.051s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.242118] env[62368]: ERROR nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Traceback (most recent call last): [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self.driver.spawn(context, instance, image_meta, [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] vm_ref = self.build_virtual_machine(instance, [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.242118] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] for vif in network_info: [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] return self._sync_wrapper(fn, *args, **kwargs) [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self.wait() [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self[:] = self._gt.wait() [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] return self._exit_event.wait() [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] result = hub.switch() [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.243110] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] return self.greenlet.switch() [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] result = function(*args, **kwargs) [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] return func(*args, **kwargs) [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] raise e [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] nwinfo = self.network_api.allocate_for_instance( [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] created_port_ids = self._update_ports_for_instance( [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] with excutils.save_and_reraise_exception(): [ 548.243774] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] self.force_reraise() [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] raise self.value [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] updated_port = self._update_port( [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] _ensure_no_port_binding_failure(port) [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] raise exception.PortBindingFailed(port_id=port['id']) [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] nova.exception.PortBindingFailed: Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. [ 548.244405] env[62368]: ERROR nova.compute.manager [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] [ 548.245626] env[62368]: DEBUG nova.compute.utils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.245626] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.841s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.245626] env[62368]: INFO nova.compute.claims [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.248465] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Build of instance 0144a1d0-bc85-4c8c-9342-8c12da4b8967 was re-scheduled: Binding failed for port 65fbf44f-a1d8-46f7-8199-6214af4ebc98, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 548.248916] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 548.249192] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Acquiring lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.249642] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Acquired lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.249642] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 548.461148] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.461426] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.461582] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.461765] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.461899] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.462064] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.462445] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.462508] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.462676] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.462829] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.463009] env[62368]: DEBUG nova.virt.hardware [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.463874] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e267f29-4add-4958-a8dc-78bc28c11c84 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.473129] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58279654-085b-43e9-92bf-af95eca0ea34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.488644] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 548.496296] env[62368]: DEBUG oslo.service.loopingcall [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.496530] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 548.496747] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58acf6cc-da26-48c0-b112-ac3a123884ee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.518173] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 548.518173] env[62368]: value = "task-1198151" [ 548.518173] env[62368]: _type = "Task" [ 548.518173] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.529675] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198151, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.784114] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.906601] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.033365] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198151, 'name': CreateVM_Task, 'duration_secs': 0.33925} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.033365] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 549.033865] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.034053] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.034461] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 549.034881] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2255555d-f628-4076-9db2-402938f73e4b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.040605] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 549.040605] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]525f4185-1f2d-a8d2-e2c0-5fc8d336b3ff" [ 549.040605] env[62368]: _type = "Task" [ 549.040605] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.054765] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525f4185-1f2d-a8d2-e2c0-5fc8d336b3ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.411486] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Releasing lock "refresh_cache-0144a1d0-bc85-4c8c-9342-8c12da4b8967" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.411779] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.412362] env[62368]: DEBUG nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.412362] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 549.439428] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.553765] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525f4185-1f2d-a8d2-e2c0-5fc8d336b3ff, 'name': SearchDatastore_Task, 'duration_secs': 0.01162} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.556184] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.556433] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 549.556910] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.556910] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.557161] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 549.558229] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6cd6d88-4c3f-46b6-b0f1-ec49ed5a20b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.567429] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 549.567429] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 549.567759] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c5b891a-f1ce-4eff-90fa-e2e0c2ff72e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.574744] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 549.574744] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52486c65-863a-fc89-c76c-06446195eaa0" [ 549.574744] env[62368]: _type = "Task" [ 549.574744] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.589575] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52486c65-863a-fc89-c76c-06446195eaa0, 'name': SearchDatastore_Task, 'duration_secs': 0.010943} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.590228] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb4c24f8-d3c5-43e1-884c-4377a20b58e3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.598689] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 549.598689] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52cc65c6-5839-b4ef-1a52-f26e38f5d4fa" [ 549.598689] env[62368]: _type = "Task" [ 549.598689] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.608306] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cc65c6-5839-b4ef-1a52-f26e38f5d4fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.637905] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9c6010-17e5-43f1-95f5-d8c2e6a10e19 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.647146] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8778605c-1cf9-49bd-8f9e-830df92912cf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.683768] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1add45ac-743e-42e6-a796-5811abe74e93 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.692581] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f83849-2808-447d-b45f-014549f22a81 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.709374] env[62368]: DEBUG nova.compute.provider_tree [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.811126] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Acquiring lock "bcecb4c6-6675-403d-9a3d-b96156b1d56a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.811470] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Lock "bcecb4c6-6675-403d-9a3d-b96156b1d56a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.941075] env[62368]: DEBUG nova.network.neutron [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.112432] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cc65c6-5839-b4ef-1a52-f26e38f5d4fa, 'name': SearchDatastore_Task, 'duration_secs': 0.009941} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.112935] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.114139] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 550.115313] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6855b19-d208-4937-9d50-1cc2e22e2da3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.125321] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 550.125321] env[62368]: value = "task-1198152" [ 550.125321] env[62368]: _type = "Task" [ 550.125321] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.135191] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.215161] env[62368]: DEBUG nova.scheduler.client.report [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.377410] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "4fde6065-41a7-4a95-96df-e442762dad24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.377652] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "4fde6065-41a7-4a95-96df-e442762dad24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.446972] env[62368]: INFO nova.compute.manager [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] [instance: 0144a1d0-bc85-4c8c-9342-8c12da4b8967] Took 1.03 seconds to deallocate network for instance. [ 550.547122] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "c9f55aeb-273b-4dae-b871-cfb1572e6535" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.547122] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "c9f55aeb-273b-4dae-b871-cfb1572e6535" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.639447] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501469} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.639726] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 550.639935] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 550.640215] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8fc70567-984a-4bd2-8bfc-dbb68cbb21f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.649624] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 550.649624] env[62368]: value = "task-1198153" [ 550.649624] env[62368]: _type = "Task" [ 550.649624] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.667226] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198153, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.720911] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.721527] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.724183] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.790s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.162611] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198153, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109252} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.162611] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 551.163840] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03bf51e-57cc-49b6-ac17-198abbe3c4bb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.190586] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 551.190943] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be174083-fcef-488a-ad28-14d6f9b47e31 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.213361] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 551.213361] env[62368]: value = "task-1198154" [ 551.213361] env[62368]: _type = "Task" [ 551.213361] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.224486] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198154, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.232195] env[62368]: DEBUG nova.compute.utils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.233827] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 551.233999] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 551.297637] env[62368]: DEBUG nova.policy [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f57279b5eaaa45be83b8026213c91099', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fc4b3b46106480693f6e18f142bb31d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.489235] env[62368]: INFO nova.scheduler.client.report [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Deleted allocations for instance 0144a1d0-bc85-4c8c-9342-8c12da4b8967 [ 551.728874] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198154, 'name': ReconfigVM_Task, 'duration_secs': 0.339025} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.729204] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175/0eb2a772-7635-481a-ae2c-9f23669dd175.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 551.730103] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92987f1a-58eb-4988-b85a-ab87c189395c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.737112] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.747066] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 551.747066] env[62368]: value = "task-1198155" [ 551.747066] env[62368]: _type = "Task" [ 551.747066] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.764503] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198155, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.774018] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8187d91f-9324-4a89-acd0-3b272d03fc8f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.780671] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22a3291-d020-4c5d-9327-fbf5e9d8f3fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.819021] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da81857b-bf78-4498-ae2b-cb1f93345610 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.828248] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c72c94-31e7-4dd9-99b2-83314fa08dfd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.843894] env[62368]: DEBUG nova.compute.provider_tree [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 552.012761] env[62368]: DEBUG oslo_concurrency.lockutils [None req-87da6cd9-33fb-41c7-a992-e19ab4a8b8b4 tempest-ServerDiagnosticsTest-1042908688 tempest-ServerDiagnosticsTest-1042908688-project-member] Lock "0144a1d0-bc85-4c8c-9342-8c12da4b8967" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.738s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.046488] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Successfully created port: e2385327-f738-43e8-bfbc-6503b85ad45e {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.262800] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198155, 'name': Rename_Task, 'duration_secs': 0.154956} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.263099] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 552.263375] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6e4bb43-a1f3-4553-b67a-f3adea76df27 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.271269] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Waiting for the task: (returnval){ [ 552.271269] env[62368]: value = "task-1198156" [ 552.271269] env[62368]: _type = "Task" [ 552.271269] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.281375] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198156, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.372446] env[62368]: ERROR nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [req-0b2dbcc0-0c26-4438-9992-ebb6be448357] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0b2dbcc0-0c26-4438-9992-ebb6be448357"}]}: nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 552.396084] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 552.415807] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 552.416294] env[62368]: DEBUG nova.compute.provider_tree [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 552.430813] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 552.453916] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 552.517526] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.761164] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.786862] env[62368]: DEBUG oslo_vmware.api [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Task: {'id': task-1198156, 'name': PowerOnVM_Task, 'duration_secs': 0.503789} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.790595] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 552.790765] env[62368]: DEBUG nova.compute.manager [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 552.793387] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f04d55-d657-4852-888e-ee68b7dced1a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.823961] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.824409] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.825206] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.825503] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.825891] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.827298] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.827298] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.827298] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.827298] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.827298] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.827472] env[62368]: DEBUG nova.virt.hardware [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.828111] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290b060a-c6cf-43c3-8d60-4f365c3f0b36 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.838851] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c298f78a-60a3-486a-98b1-6c5e8fdc6a00 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.975894] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d44611-5938-422a-93ce-4bf03b7ac162 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.987072] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956f4a74-f986-48d4-bd3a-ca58f02fed78 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.025187] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc1f2b5-0c0c-461f-8d01-1b77f07d276e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.036517] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277ab52e-80b4-4a35-841e-d437d8027cf3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.054992] env[62368]: DEBUG nova.compute.provider_tree [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 553.057203] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.327061] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.595849] env[62368]: ERROR nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [req-119ec5c4-b2d3-4d0c-ab6b-fe9e7fa99aa7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-119ec5c4-b2d3-4d0c-ab6b-fe9e7fa99aa7"}]}: nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 553.614739] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 553.631512] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 553.631722] env[62368]: DEBUG nova.compute.provider_tree [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 553.651043] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 553.671731] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 554.055945] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "9cdf70bc-0b9b-44fa-80e7-3584fc104959" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.055945] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "9cdf70bc-0b9b-44fa-80e7-3584fc104959" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.127657] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "8e6332f4-2b1e-404d-aa43-a13a72edd044" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.127885] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "8e6332f4-2b1e-404d-aa43-a13a72edd044" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.290939] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236f1ffb-f5bf-494b-a3fc-7b5b1ed8fc68 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.303942] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d882e70d-3430-4980-8bb7-5c671e1a18a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.340550] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2217bc95-20d8-4c32-ad7d-4a3bd325b86c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.349591] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f765f918-d48f-4a94-905a-c52afb1e61b2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.363352] env[62368]: DEBUG nova.compute.provider_tree [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 554.593506] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "0eb2a772-7635-481a-ae2c-9f23669dd175" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.594281] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "0eb2a772-7635-481a-ae2c-9f23669dd175" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.594943] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "0eb2a772-7635-481a-ae2c-9f23669dd175-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.595816] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "0eb2a772-7635-481a-ae2c-9f23669dd175-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.597169] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "0eb2a772-7635-481a-ae2c-9f23669dd175-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.601788] env[62368]: INFO nova.compute.manager [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Terminating instance [ 554.603919] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "refresh_cache-0eb2a772-7635-481a-ae2c-9f23669dd175" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.604561] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquired lock "refresh_cache-0eb2a772-7635-481a-ae2c-9f23669dd175" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.604775] env[62368]: DEBUG nova.network.neutron [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 554.919247] env[62368]: DEBUG nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 37 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 554.921258] env[62368]: DEBUG nova.compute.provider_tree [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 37 to 38 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 554.921258] env[62368]: DEBUG nova.compute.provider_tree [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 554.950201] env[62368]: DEBUG nova.compute.manager [req-16ba89df-234c-4d4f-9adc-5133029c8fbf req-5944cda2-a73b-4166-b4a2-63a690f5a308 service nova] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Received event network-changed-e2385327-f738-43e8-bfbc-6503b85ad45e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 554.950600] env[62368]: DEBUG nova.compute.manager [req-16ba89df-234c-4d4f-9adc-5133029c8fbf req-5944cda2-a73b-4166-b4a2-63a690f5a308 service nova] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Refreshing instance network info cache due to event network-changed-e2385327-f738-43e8-bfbc-6503b85ad45e. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 554.952558] env[62368]: DEBUG oslo_concurrency.lockutils [req-16ba89df-234c-4d4f-9adc-5133029c8fbf req-5944cda2-a73b-4166-b4a2-63a690f5a308 service nova] Acquiring lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.952558] env[62368]: DEBUG oslo_concurrency.lockutils [req-16ba89df-234c-4d4f-9adc-5133029c8fbf req-5944cda2-a73b-4166-b4a2-63a690f5a308 service nova] Acquired lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.952558] env[62368]: DEBUG nova.network.neutron [req-16ba89df-234c-4d4f-9adc-5133029c8fbf req-5944cda2-a73b-4166-b4a2-63a690f5a308 service nova] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Refreshing network info cache for port e2385327-f738-43e8-bfbc-6503b85ad45e {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 555.139759] env[62368]: DEBUG nova.network.neutron [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.146819] env[62368]: ERROR nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. [ 555.146819] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 555.146819] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.146819] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 555.146819] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 555.146819] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 555.146819] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 555.146819] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 555.146819] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.146819] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 555.146819] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.146819] env[62368]: ERROR nova.compute.manager raise self.value [ 555.146819] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 555.146819] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 555.146819] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.146819] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 555.147817] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.147817] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 555.147817] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. [ 555.147817] env[62368]: ERROR nova.compute.manager [ 555.147817] env[62368]: Traceback (most recent call last): [ 555.147817] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 555.147817] env[62368]: listener.cb(fileno) [ 555.147817] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.147817] env[62368]: result = function(*args, **kwargs) [ 555.147817] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 555.147817] env[62368]: return func(*args, **kwargs) [ 555.147817] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.147817] env[62368]: raise e [ 555.147817] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.147817] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 555.147817] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 555.147817] env[62368]: created_port_ids = self._update_ports_for_instance( [ 555.147817] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 555.147817] env[62368]: with excutils.save_and_reraise_exception(): [ 555.147817] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.147817] env[62368]: self.force_reraise() [ 555.147817] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.147817] env[62368]: raise self.value [ 555.147817] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 555.147817] env[62368]: updated_port = self._update_port( [ 555.147817] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.147817] env[62368]: _ensure_no_port_binding_failure(port) [ 555.147817] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.147817] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 555.148976] env[62368]: nova.exception.PortBindingFailed: Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. [ 555.148976] env[62368]: Removing descriptor: 15 [ 555.148976] env[62368]: ERROR nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Traceback (most recent call last): [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] yield resources [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self.driver.spawn(context, instance, image_meta, [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.148976] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] vm_ref = self.build_virtual_machine(instance, [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] for vif in network_info: [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] return self._sync_wrapper(fn, *args, **kwargs) [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self.wait() [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self[:] = self._gt.wait() [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] return self._exit_event.wait() [ 555.149355] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] result = hub.switch() [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] return self.greenlet.switch() [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] result = function(*args, **kwargs) [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] return func(*args, **kwargs) [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] raise e [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] nwinfo = self.network_api.allocate_for_instance( [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 555.149774] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] created_port_ids = self._update_ports_for_instance( [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] with excutils.save_and_reraise_exception(): [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self.force_reraise() [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] raise self.value [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] updated_port = self._update_port( [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] _ensure_no_port_binding_failure(port) [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.150441] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] raise exception.PortBindingFailed(port_id=port['id']) [ 555.150799] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] nova.exception.PortBindingFailed: Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. [ 555.150799] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] [ 555.150799] env[62368]: INFO nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Terminating instance [ 555.150799] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.428891] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 4.705s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.429581] env[62368]: ERROR nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Traceback (most recent call last): [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self.driver.spawn(context, instance, image_meta, [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] vm_ref = self.build_virtual_machine(instance, [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.429581] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] for vif in network_info: [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] return self._sync_wrapper(fn, *args, **kwargs) [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self.wait() [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self[:] = self._gt.wait() [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] return self._exit_event.wait() [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] result = hub.switch() [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.430191] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] return self.greenlet.switch() [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] result = function(*args, **kwargs) [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] return func(*args, **kwargs) [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] raise e [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] nwinfo = self.network_api.allocate_for_instance( [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] created_port_ids = self._update_ports_for_instance( [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] with excutils.save_and_reraise_exception(): [ 555.432166] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] self.force_reraise() [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] raise self.value [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] updated_port = self._update_port( [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] _ensure_no_port_binding_failure(port) [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] raise exception.PortBindingFailed(port_id=port['id']) [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] nova.exception.PortBindingFailed: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. [ 555.435515] env[62368]: ERROR nova.compute.manager [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] [ 555.436273] env[62368]: DEBUG nova.compute.utils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 555.436273] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.771s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.438646] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Build of instance 50feff45-9ede-4d4d-b136-bfb5fd916dc9 was re-scheduled: Binding failed for port 4b89762c-af45-4c1d-aea6-e6058910e094, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 555.439124] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 555.439365] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Acquiring lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.439513] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Acquired lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.439669] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 555.451613] env[62368]: DEBUG nova.network.neutron [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.491220] env[62368]: DEBUG nova.network.neutron [req-16ba89df-234c-4d4f-9adc-5133029c8fbf req-5944cda2-a73b-4166-b4a2-63a690f5a308 service nova] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.729646] env[62368]: DEBUG nova.network.neutron [req-16ba89df-234c-4d4f-9adc-5133029c8fbf req-5944cda2-a73b-4166-b4a2-63a690f5a308 service nova] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.954642] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Releasing lock "refresh_cache-0eb2a772-7635-481a-ae2c-9f23669dd175" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.956026] env[62368]: DEBUG nova.compute.manager [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 555.956026] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 555.956876] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a42666-228c-49e3-9561-be908d423944 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.970185] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 555.970476] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2ebe85b-36cd-4ddc-9ba4-f5ac264d5f61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.979266] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.984862] env[62368]: DEBUG oslo_vmware.api [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 555.984862] env[62368]: value = "task-1198157" [ 555.984862] env[62368]: _type = "Task" [ 555.984862] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.999901] env[62368]: DEBUG oslo_vmware.api [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.146964] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.239313] env[62368]: DEBUG oslo_concurrency.lockutils [req-16ba89df-234c-4d4f-9adc-5133029c8fbf req-5944cda2-a73b-4166-b4a2-63a690f5a308 service nova] Releasing lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.241315] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquired lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.241549] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.438842] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b43d60-64a2-4dad-8b7e-649d0d2540cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.452018] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459f3775-4891-43b1-9a6f-db4053f6ba69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.486036] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bae8d5-a2e2-4211-9a0b-907f922903d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.498835] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164389d8-cad9-4942-92f0-5be0c1e49614 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.503752] env[62368]: DEBUG oslo_vmware.api [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198157, 'name': PowerOffVM_Task, 'duration_secs': 0.210506} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.503752] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 556.503752] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 556.504249] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d8d870b-fb01-49e6-8204-5b147ccccd99 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.513735] env[62368]: DEBUG nova.compute.provider_tree [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.520436] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "09ac5c26-9842-440c-bc14-a84c2eb23cf5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.520436] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "09ac5c26-9842-440c-bc14-a84c2eb23cf5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.540107] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 556.540339] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 556.540521] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleting the datastore file [datastore1] 0eb2a772-7635-481a-ae2c-9f23669dd175 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 556.540777] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe9f42ce-231e-4a89-9da4-1050a23d7194 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.546974] env[62368]: DEBUG oslo_vmware.api [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for the task: (returnval){ [ 556.546974] env[62368]: value = "task-1198159" [ 556.546974] env[62368]: _type = "Task" [ 556.546974] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.556124] env[62368]: DEBUG oslo_vmware.api [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.650635] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Releasing lock "refresh_cache-50feff45-9ede-4d4d-b136-bfb5fd916dc9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.650635] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 556.650809] env[62368]: DEBUG nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.651993] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 556.672063] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.768944] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.849563] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.021675] env[62368]: DEBUG nova.scheduler.client.report [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.057617] env[62368]: DEBUG oslo_vmware.api [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Task: {'id': task-1198159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100664} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.059496] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 557.059726] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 557.060035] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 557.060225] env[62368]: INFO nova.compute.manager [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Took 1.10 seconds to destroy the instance on the hypervisor. [ 557.060467] env[62368]: DEBUG oslo.service.loopingcall [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.060657] env[62368]: DEBUG nova.compute.manager [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.060749] env[62368]: DEBUG nova.network.neutron [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 557.078384] env[62368]: DEBUG nova.network.neutron [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.167798] env[62368]: DEBUG nova.compute.manager [req-10c9dfe2-43fa-409f-a39b-5556b736425c req-73a1ee19-3438-4ab9-b72b-67eb052564bf service nova] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Received event network-vif-deleted-e2385327-f738-43e8-bfbc-6503b85ad45e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.175256] env[62368]: DEBUG nova.network.neutron [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.353189] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Releasing lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.353611] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.353811] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 557.354166] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d15ad08b-d64f-4ece-b822-b3738891f679 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.364140] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04a8e16-b3ab-4326-b4a9-2e4b8c5a35ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.386715] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 54e3c531-9253-4fc0-999c-2cb9a8388d08 could not be found. [ 557.386920] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 557.387385] env[62368]: INFO nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Took 0.03 seconds to destroy the instance on the hypervisor. [ 557.387385] env[62368]: DEBUG oslo.service.loopingcall [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.387572] env[62368]: DEBUG nova.compute.manager [-] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.387667] env[62368]: DEBUG nova.network.neutron [-] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 557.404821] env[62368]: DEBUG nova.network.neutron [-] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.528581] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.095s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.530809] env[62368]: ERROR nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Traceback (most recent call last): [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self.driver.spawn(context, instance, image_meta, [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] vm_ref = self.build_virtual_machine(instance, [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.530809] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] for vif in network_info: [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] return self._sync_wrapper(fn, *args, **kwargs) [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self.wait() [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self[:] = self._gt.wait() [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] return self._exit_event.wait() [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] result = hub.switch() [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.531254] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] return self.greenlet.switch() [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] result = function(*args, **kwargs) [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] return func(*args, **kwargs) [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] raise e [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] nwinfo = self.network_api.allocate_for_instance( [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] created_port_ids = self._update_ports_for_instance( [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] with excutils.save_and_reraise_exception(): [ 557.531633] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] self.force_reraise() [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] raise self.value [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] updated_port = self._update_port( [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] _ensure_no_port_binding_failure(port) [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] raise exception.PortBindingFailed(port_id=port['id']) [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] nova.exception.PortBindingFailed: Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. [ 557.532020] env[62368]: ERROR nova.compute.manager [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] [ 557.532435] env[62368]: DEBUG nova.compute.utils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.533618] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.169s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.538121] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Build of instance 5d0e30b7-17b1-46c1-9cd0-05df00c9923e was re-scheduled: Binding failed for port fa4e60cb-8bd8-498c-b401-8f7ee20a4abd, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.538958] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.539359] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Acquiring lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.539629] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Acquired lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.539856] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 557.582660] env[62368]: DEBUG nova.network.neutron [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.679918] env[62368]: INFO nova.compute.manager [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] [instance: 50feff45-9ede-4d4d-b136-bfb5fd916dc9] Took 1.03 seconds to deallocate network for instance. [ 557.911419] env[62368]: DEBUG nova.network.neutron [-] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.067638] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.086461] env[62368]: INFO nova.compute.manager [-] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Took 1.03 seconds to deallocate network for instance. [ 558.197184] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.414548] env[62368]: INFO nova.compute.manager [-] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Took 1.03 seconds to deallocate network for instance. [ 558.419115] env[62368]: DEBUG nova.compute.claims [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 558.419318] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.544811] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a5a441-9644-49a8-a9ec-7c658e6197e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.554027] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64ca819-fbca-4c0f-9e71-25e23c6455d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.592034] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f0015c-c541-4040-942a-3727450d6283 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.601207] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f1478c-ad3e-46b1-aaf9-131e51ee765e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.603593] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.618266] env[62368]: DEBUG nova.compute.provider_tree [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.701903] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Releasing lock "refresh_cache-5d0e30b7-17b1-46c1-9cd0-05df00c9923e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.702162] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 558.702354] env[62368]: DEBUG nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.702530] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 558.717065] env[62368]: INFO nova.scheduler.client.report [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Deleted allocations for instance 50feff45-9ede-4d4d-b136-bfb5fd916dc9 [ 558.722755] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.122437] env[62368]: DEBUG nova.scheduler.client.report [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.142255] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Acquiring lock "9518175e-e4ac-42fc-bd90-cc7e204ef2e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.142502] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Lock "9518175e-e4ac-42fc-bd90-cc7e204ef2e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.226809] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a6be504-dbfd-4c7b-8521-bf12c5e830bf tempest-ServerExternalEventsTest-991779733 tempest-ServerExternalEventsTest-991779733-project-member] Lock "50feff45-9ede-4d4d-b136-bfb5fd916dc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.124s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.227438] env[62368]: DEBUG nova.network.neutron [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.630854] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.097s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.630854] env[62368]: ERROR nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. [ 559.630854] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Traceback (most recent call last): [ 559.630854] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.630854] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self.driver.spawn(context, instance, image_meta, [ 559.630854] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.630854] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.630854] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.630854] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] vm_ref = self.build_virtual_machine(instance, [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] for vif in network_info: [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] return self._sync_wrapper(fn, *args, **kwargs) [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self.wait() [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self[:] = self._gt.wait() [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] return self._exit_event.wait() [ 559.631513] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] result = hub.switch() [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] return self.greenlet.switch() [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] result = function(*args, **kwargs) [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] return func(*args, **kwargs) [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] raise e [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] nwinfo = self.network_api.allocate_for_instance( [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.631931] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] created_port_ids = self._update_ports_for_instance( [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] with excutils.save_and_reraise_exception(): [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] self.force_reraise() [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] raise self.value [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] updated_port = self._update_port( [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] _ensure_no_port_binding_failure(port) [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.632463] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] raise exception.PortBindingFailed(port_id=port['id']) [ 559.632852] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] nova.exception.PortBindingFailed: Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. [ 559.632852] env[62368]: ERROR nova.compute.manager [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] [ 559.632852] env[62368]: DEBUG nova.compute.utils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 559.632852] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.893s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.635479] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Build of instance 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4 was re-scheduled: Binding failed for port 8ec1919a-6e92-4c70-b8b0-9f3b01caeabf, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 559.636657] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 559.636657] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Acquiring lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.636657] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Acquired lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.636657] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 559.735202] env[62368]: INFO nova.compute.manager [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] [instance: 5d0e30b7-17b1-46c1-9cd0-05df00c9923e] Took 1.03 seconds to deallocate network for instance. [ 559.739550] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 559.991809] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.992100] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.165095] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.265535] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.294572] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.573154] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57e3fe6-6618-4455-bf88-4383d451c231 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.583141] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bdea47-2e82-4fe9-af08-a95562d3927a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.614911] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8d9426-59c3-4dca-88fa-76b1756ea3bb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.622419] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4879e3a1-611d-46df-b013-9afa1f1488ac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.640419] env[62368]: DEBUG nova.compute.provider_tree [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.666229] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "49a8e94b-b167-41d1-b748-1a2fa4e2b940" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.666526] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "49a8e94b-b167-41d1-b748-1a2fa4e2b940" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.786017] env[62368]: INFO nova.scheduler.client.report [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Deleted allocations for instance 5d0e30b7-17b1-46c1-9cd0-05df00c9923e [ 560.796297] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Releasing lock "refresh_cache-431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.796531] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 560.796794] env[62368]: DEBUG nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.797007] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.828053] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.143618] env[62368]: DEBUG nova.scheduler.client.report [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.295091] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66752e2c-8eb5-4c64-ab7b-7900fd7c5b89 tempest-ServersAdminNegativeTestJSON-1140433656 tempest-ServersAdminNegativeTestJSON-1140433656-project-member] Lock "5d0e30b7-17b1-46c1-9cd0-05df00c9923e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.260s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.331215] env[62368]: DEBUG nova.network.neutron [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.648897] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.016s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.649749] env[62368]: ERROR nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Traceback (most recent call last): [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self.driver.spawn(context, instance, image_meta, [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] vm_ref = self.build_virtual_machine(instance, [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.649749] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] for vif in network_info: [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] return self._sync_wrapper(fn, *args, **kwargs) [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self.wait() [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self[:] = self._gt.wait() [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] return self._exit_event.wait() [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] result = hub.switch() [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.650391] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] return self.greenlet.switch() [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] result = function(*args, **kwargs) [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] return func(*args, **kwargs) [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] raise e [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] nwinfo = self.network_api.allocate_for_instance( [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] created_port_ids = self._update_ports_for_instance( [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] with excutils.save_and_reraise_exception(): [ 561.650819] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] self.force_reraise() [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] raise self.value [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] updated_port = self._update_port( [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] _ensure_no_port_binding_failure(port) [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] raise exception.PortBindingFailed(port_id=port['id']) [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] nova.exception.PortBindingFailed: Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. [ 561.651365] env[62368]: ERROR nova.compute.manager [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] [ 561.651687] env[62368]: DEBUG nova.compute.utils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 561.652508] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.299s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.655218] env[62368]: INFO nova.compute.claims [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.660550] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Build of instance 3356d528-14ea-4a7d-acef-6c6f56195f31 was re-scheduled: Binding failed for port d03daeea-1059-48e6-bb5e-3e0939733e47, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 561.661049] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 561.662139] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.662139] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquired lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.662139] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.802180] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.833604] env[62368]: INFO nova.compute.manager [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] [instance: 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4] Took 1.04 seconds to deallocate network for instance. [ 562.196330] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.323885] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.485263] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.865426] env[62368]: INFO nova.scheduler.client.report [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Deleted allocations for instance 431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4 [ 562.990580] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Releasing lock "refresh_cache-3356d528-14ea-4a7d-acef-6c6f56195f31" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.990865] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 562.991298] env[62368]: DEBUG nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.991472] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 563.007704] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.131460] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467a9534-a500-40a2-a2ec-8d1a81100109 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.139155] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eeb63ad-436c-4356-9d8c-4c5962b1a991 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.175114] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909e33cd-4440-44f0-a7bb-15e8c7d5ce52 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.181731] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d527742c-a425-4b1d-86e7-0674948ea693 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.196191] env[62368]: DEBUG nova.compute.provider_tree [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.378560] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8a5e024-32da-4cd9-a020-89e16e8f09e8 tempest-ServersTestFqdnHostnames-1554111744 tempest-ServersTestFqdnHostnames-1554111744-project-member] Lock "431d3ce0-a5ed-46c0-9c94-9a8d2fbdc4f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.023s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.510972] env[62368]: DEBUG nova.network.neutron [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.701164] env[62368]: DEBUG nova.scheduler.client.report [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.881229] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 564.015533] env[62368]: INFO nova.compute.manager [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 3356d528-14ea-4a7d-acef-6c6f56195f31] Took 1.02 seconds to deallocate network for instance. [ 564.209961] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.555s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.209961] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 564.211644] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.914s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.410160] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.720163] env[62368]: DEBUG nova.compute.utils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 564.720163] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 564.720163] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 564.838322] env[62368]: DEBUG nova.policy [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0422f0f3fa0c480f8e9daf6beca77a8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e6890a02270457d9185f9e384454cd4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 565.070316] env[62368]: INFO nova.scheduler.client.report [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Deleted allocations for instance 3356d528-14ea-4a7d-acef-6c6f56195f31 [ 565.191766] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332f6d0c-b254-4d38-8c48-f4fba820e23d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.201992] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb285089-8a8e-4de4-9cb0-e09e7858500a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.235810] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.239437] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ea3286-e2d0-482c-a8cb-3bd0e939ec40 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.247483] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffa85cc-4b12-4755-8543-d05aeef589d1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.260803] env[62368]: DEBUG nova.compute.provider_tree [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.582147] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eff12c0b-a177-4e3f-9b54-028e77f1c7f7 tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "3356d528-14ea-4a7d-acef-6c6f56195f31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.157s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.750298] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Successfully created port: 59966839-e7d4-4c37-a1ba-018a59b8d38e {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 565.763524] env[62368]: DEBUG nova.scheduler.client.report [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.084182] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 566.249073] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.271259] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.056s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.271259] env[62368]: ERROR nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. [ 566.271259] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Traceback (most recent call last): [ 566.271259] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.271259] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self.driver.spawn(context, instance, image_meta, [ 566.271259] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 566.271259] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.271259] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.271259] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] vm_ref = self.build_virtual_machine(instance, [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] for vif in network_info: [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] return self._sync_wrapper(fn, *args, **kwargs) [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self.wait() [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self[:] = self._gt.wait() [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] return self._exit_event.wait() [ 566.271584] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] result = hub.switch() [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] return self.greenlet.switch() [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] result = function(*args, **kwargs) [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] return func(*args, **kwargs) [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] raise e [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] nwinfo = self.network_api.allocate_for_instance( [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.271954] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] created_port_ids = self._update_ports_for_instance( [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] with excutils.save_and_reraise_exception(): [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] self.force_reraise() [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] raise self.value [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] updated_port = self._update_port( [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] _ensure_no_port_binding_failure(port) [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.272346] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] raise exception.PortBindingFailed(port_id=port['id']) [ 566.272681] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] nova.exception.PortBindingFailed: Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. [ 566.272681] env[62368]: ERROR nova.compute.manager [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] [ 566.272681] env[62368]: DEBUG nova.compute.utils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 566.272879] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.338s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.276377] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Build of instance 73c833ef-7761-4372-9d02-b6dfc945e415 was re-scheduled: Binding failed for port 4e18f0de-825f-466a-80b7-7bb7a8333db5, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 566.276697] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 566.281983] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Acquiring lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.281983] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Acquired lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.281983] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.286135] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.286135] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.286135] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.286507] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.286507] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.286507] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.286507] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.286638] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.286704] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.286860] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.287035] env[62368]: DEBUG nova.virt.hardware [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.287883] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d45bd2-661a-4255-9a98-84a924c5a330 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.297976] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b3460d-2498-45a0-b71c-e76263a94171 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.612677] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.838412] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.955300] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.233266] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc9f54d-375b-44ba-9f31-6b7281b9f995 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.241422] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0201bbe5-bc82-4ac8-9eb7-d5f5db001223 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.277988] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd7d0f9-131c-444e-a922-98b37ac7c148 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.286183] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e4e209-bf74-4601-8cb0-b1bd4506fbff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.300734] env[62368]: DEBUG nova.compute.provider_tree [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.460587] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Releasing lock "refresh_cache-73c833ef-7761-4372-9d02-b6dfc945e415" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.460746] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 567.462937] env[62368]: DEBUG nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.462937] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.482678] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.809228] env[62368]: DEBUG nova.scheduler.client.report [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.987190] env[62368]: DEBUG nova.network.neutron [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.313183] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.313183] env[62368]: ERROR nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. [ 568.313183] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Traceback (most recent call last): [ 568.313183] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.313183] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self.driver.spawn(context, instance, image_meta, [ 568.313183] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.313183] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.313183] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.313183] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] vm_ref = self.build_virtual_machine(instance, [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] for vif in network_info: [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] return self._sync_wrapper(fn, *args, **kwargs) [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self.wait() [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self[:] = self._gt.wait() [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] return self._exit_event.wait() [ 568.313572] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] result = hub.switch() [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] return self.greenlet.switch() [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] result = function(*args, **kwargs) [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] return func(*args, **kwargs) [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] raise e [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] nwinfo = self.network_api.allocate_for_instance( [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.313969] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] created_port_ids = self._update_ports_for_instance( [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] with excutils.save_and_reraise_exception(): [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] self.force_reraise() [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] raise self.value [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] updated_port = self._update_port( [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] _ensure_no_port_binding_failure(port) [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.314340] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] raise exception.PortBindingFailed(port_id=port['id']) [ 568.314702] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] nova.exception.PortBindingFailed: Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. [ 568.314702] env[62368]: ERROR nova.compute.manager [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] [ 568.314702] env[62368]: DEBUG nova.compute.utils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 568.316032] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Build of instance 92d0e65f-71c9-4ab8-8994-7425306b92bb was re-scheduled: Binding failed for port 5c4de9fa-fd2c-4559-b912-cfc173677a4b, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 568.316402] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 568.316679] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.316857] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquired lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.317168] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.318905] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.318s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.321200] env[62368]: INFO nova.compute.claims [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.490148] env[62368]: INFO nova.compute.manager [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] [instance: 73c833ef-7761-4372-9d02-b6dfc945e415] Took 1.03 seconds to deallocate network for instance. [ 568.860148] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.949796] env[62368]: DEBUG nova.compute.manager [req-cbb29e57-04b2-498b-aff9-d76bf69da4b3 req-6d5e8d47-3021-4473-8af1-c397ff070cbd service nova] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Received event network-changed-59966839-e7d4-4c37-a1ba-018a59b8d38e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.949962] env[62368]: DEBUG nova.compute.manager [req-cbb29e57-04b2-498b-aff9-d76bf69da4b3 req-6d5e8d47-3021-4473-8af1-c397ff070cbd service nova] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Refreshing instance network info cache due to event network-changed-59966839-e7d4-4c37-a1ba-018a59b8d38e. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 568.950466] env[62368]: DEBUG oslo_concurrency.lockutils [req-cbb29e57-04b2-498b-aff9-d76bf69da4b3 req-6d5e8d47-3021-4473-8af1-c397ff070cbd service nova] Acquiring lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.950466] env[62368]: DEBUG oslo_concurrency.lockutils [req-cbb29e57-04b2-498b-aff9-d76bf69da4b3 req-6d5e8d47-3021-4473-8af1-c397ff070cbd service nova] Acquired lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.950635] env[62368]: DEBUG nova.network.neutron [req-cbb29e57-04b2-498b-aff9-d76bf69da4b3 req-6d5e8d47-3021-4473-8af1-c397ff070cbd service nova] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Refreshing network info cache for port 59966839-e7d4-4c37-a1ba-018a59b8d38e {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 569.007765] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.481520] env[62368]: DEBUG nova.network.neutron [req-cbb29e57-04b2-498b-aff9-d76bf69da4b3 req-6d5e8d47-3021-4473-8af1-c397ff070cbd service nova] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.511591] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Releasing lock "refresh_cache-92d0e65f-71c9-4ab8-8994-7425306b92bb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.511852] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 569.512089] env[62368]: DEBUG nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 569.512261] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 569.526228] env[62368]: INFO nova.scheduler.client.report [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Deleted allocations for instance 73c833ef-7761-4372-9d02-b6dfc945e415 [ 569.552188] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.559399] env[62368]: ERROR nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. [ 569.559399] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.559399] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.559399] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.559399] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.559399] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.559399] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.559399] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.559399] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.559399] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 569.559399] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.559399] env[62368]: ERROR nova.compute.manager raise self.value [ 569.559399] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.559399] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.559399] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.559399] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.560028] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.560028] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.560028] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. [ 569.560028] env[62368]: ERROR nova.compute.manager [ 569.560028] env[62368]: Traceback (most recent call last): [ 569.560028] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.560028] env[62368]: listener.cb(fileno) [ 569.560028] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.560028] env[62368]: result = function(*args, **kwargs) [ 569.560028] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.560028] env[62368]: return func(*args, **kwargs) [ 569.560028] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.560028] env[62368]: raise e [ 569.560028] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.560028] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 569.560028] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.560028] env[62368]: created_port_ids = self._update_ports_for_instance( [ 569.560028] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.560028] env[62368]: with excutils.save_and_reraise_exception(): [ 569.560028] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.560028] env[62368]: self.force_reraise() [ 569.560028] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.560028] env[62368]: raise self.value [ 569.560028] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.560028] env[62368]: updated_port = self._update_port( [ 569.560028] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.560028] env[62368]: _ensure_no_port_binding_failure(port) [ 569.560028] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.560028] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.560825] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. [ 569.560825] env[62368]: Removing descriptor: 18 [ 569.560825] env[62368]: ERROR nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Traceback (most recent call last): [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] yield resources [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self.driver.spawn(context, instance, image_meta, [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.560825] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] vm_ref = self.build_virtual_machine(instance, [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] for vif in network_info: [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] return self._sync_wrapper(fn, *args, **kwargs) [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self.wait() [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self[:] = self._gt.wait() [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] return self._exit_event.wait() [ 569.561291] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] result = hub.switch() [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] return self.greenlet.switch() [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] result = function(*args, **kwargs) [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] return func(*args, **kwargs) [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] raise e [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] nwinfo = self.network_api.allocate_for_instance( [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.561779] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] created_port_ids = self._update_ports_for_instance( [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] with excutils.save_and_reraise_exception(): [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self.force_reraise() [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] raise self.value [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] updated_port = self._update_port( [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] _ensure_no_port_binding_failure(port) [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.562161] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] raise exception.PortBindingFailed(port_id=port['id']) [ 569.562505] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] nova.exception.PortBindingFailed: Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. [ 569.562505] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] [ 569.562505] env[62368]: INFO nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Terminating instance [ 569.566271] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Acquiring lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.696818] env[62368]: DEBUG nova.network.neutron [req-cbb29e57-04b2-498b-aff9-d76bf69da4b3 req-6d5e8d47-3021-4473-8af1-c397ff070cbd service nova] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.841441] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56412f1-edd7-41fe-94ae-db4df08a6788 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.850058] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0354893c-f058-4d61-94d5-09438c042ac9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.885374] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b827147-7407-4f50-be5a-8bd92631f18b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.895154] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f9a10d-7c7c-4d33-a9e0-ef0e59a05641 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.911287] env[62368]: DEBUG nova.compute.provider_tree [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.037032] env[62368]: DEBUG oslo_concurrency.lockutils [None req-076e44b4-a5c9-4277-898b-424b2e2c68c6 tempest-ServerDiagnosticsNegativeTest-1998148705 tempest-ServerDiagnosticsNegativeTest-1998148705-project-member] Lock "73c833ef-7761-4372-9d02-b6dfc945e415" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.149s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.061109] env[62368]: DEBUG nova.network.neutron [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.201709] env[62368]: DEBUG oslo_concurrency.lockutils [req-cbb29e57-04b2-498b-aff9-d76bf69da4b3 req-6d5e8d47-3021-4473-8af1-c397ff070cbd service nova] Releasing lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.202990] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Acquired lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.203306] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 570.417095] env[62368]: DEBUG nova.scheduler.client.report [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.539775] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 570.565344] env[62368]: INFO nova.compute.manager [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 92d0e65f-71c9-4ab8-8994-7425306b92bb] Took 1.05 seconds to deallocate network for instance. [ 570.851305] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.926742] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.608s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.927301] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.930484] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.571s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.930484] env[62368]: DEBUG nova.objects.instance [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 571.084202] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.089432] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.160017] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Acquiring lock "c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.160017] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Lock "c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.436192] env[62368]: DEBUG nova.compute.utils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.437366] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 571.437581] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 571.543072] env[62368]: DEBUG nova.policy [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f14ca41740c4efc9252097f9ffefea8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eac3ef24ae3e4e34855df574cc09969b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.550984] env[62368]: DEBUG nova.compute.manager [req-0fa27bb0-092c-41d2-951b-2d226d8d5358 req-5a769e44-1ac3-4b63-b5e7-5ae30ea55c05 service nova] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Received event network-vif-deleted-59966839-e7d4-4c37-a1ba-018a59b8d38e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 571.593497] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Releasing lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.595152] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 571.595370] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 571.595696] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1510f699-f3c9-475b-b472-d0be41e90fc8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.609601] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6476d80a-1d95-4699-ba93-e0ad2ad21a2e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.631234] env[62368]: INFO nova.scheduler.client.report [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Deleted allocations for instance 92d0e65f-71c9-4ab8-8994-7425306b92bb [ 571.648315] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eac92d1f-d160-4e06-a71c-d064acdffb19 could not be found. [ 571.648718] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 571.649012] env[62368]: INFO nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Took 0.05 seconds to destroy the instance on the hypervisor. [ 571.652037] env[62368]: DEBUG oslo.service.loopingcall [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.652037] env[62368]: DEBUG nova.compute.manager [-] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.652037] env[62368]: DEBUG nova.network.neutron [-] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.675556] env[62368]: DEBUG nova.network.neutron [-] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.942066] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.952019] env[62368]: DEBUG oslo_concurrency.lockutils [None req-862a3609-81d4-4d1d-8221-4364c31af47b tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.952123] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.895s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.955267] env[62368]: INFO nova.compute.claims [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.098105] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Acquiring lock "70055136-b08b-4654-aca7-8280158d94da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.098105] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Lock "70055136-b08b-4654-aca7-8280158d94da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.140825] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c5bf47b-e0af-4c8c-b0ab-03e9e9b5d9fa tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "92d0e65f-71c9-4ab8-8994-7425306b92bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.372s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.169104] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Successfully created port: eade0b39-4a3a-4903-9ded-828f9f9ccc95 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.177960] env[62368]: DEBUG nova.network.neutron [-] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.646615] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.680573] env[62368]: INFO nova.compute.manager [-] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Took 1.03 seconds to deallocate network for instance. [ 572.685276] env[62368]: DEBUG nova.compute.claims [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.685780] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.710186] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Acquiring lock "5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.710547] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Lock "5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.960394] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.998740] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.999269] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.999269] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.999468] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 573.000373] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 573.000373] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 573.000373] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 573.000373] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 573.000373] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 573.000571] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 573.001196] env[62368]: DEBUG nova.virt.hardware [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 573.002198] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c40169-9711-4741-8643-75aa070e7421 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.014455] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad16efe5-3a11-4ae5-a0e9-b0b0ca8a6fc0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.169859] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.440325] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96def8a-5bd6-4d51-8a7d-a081ea9cb488 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.450437] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292b3084-01a1-4ecd-83f8-82223a1641f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.489310] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5420388f-a735-4b13-861a-fdb48e03cb65 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.498113] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8501e3a-8e3f-4523-b109-b8d0a72baa02 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.515366] env[62368]: DEBUG nova.compute.provider_tree [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.019136] env[62368]: DEBUG nova.scheduler.client.report [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.160981] env[62368]: DEBUG nova.compute.manager [req-5cbb64d8-6e1b-4474-9bff-912974d69e70 req-bcf1b4b3-c113-454d-9a47-224d343e3dc9 service nova] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Received event network-changed-eade0b39-4a3a-4903-9ded-828f9f9ccc95 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.162173] env[62368]: DEBUG nova.compute.manager [req-5cbb64d8-6e1b-4474-9bff-912974d69e70 req-bcf1b4b3-c113-454d-9a47-224d343e3dc9 service nova] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Refreshing instance network info cache due to event network-changed-eade0b39-4a3a-4903-9ded-828f9f9ccc95. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 574.162575] env[62368]: DEBUG oslo_concurrency.lockutils [req-5cbb64d8-6e1b-4474-9bff-912974d69e70 req-bcf1b4b3-c113-454d-9a47-224d343e3dc9 service nova] Acquiring lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.162638] env[62368]: DEBUG oslo_concurrency.lockutils [req-5cbb64d8-6e1b-4474-9bff-912974d69e70 req-bcf1b4b3-c113-454d-9a47-224d343e3dc9 service nova] Acquired lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.162857] env[62368]: DEBUG nova.network.neutron [req-5cbb64d8-6e1b-4474-9bff-912974d69e70 req-bcf1b4b3-c113-454d-9a47-224d343e3dc9 service nova] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Refreshing network info cache for port eade0b39-4a3a-4903-9ded-828f9f9ccc95 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 574.318479] env[62368]: ERROR nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. [ 574.318479] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 574.318479] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.318479] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 574.318479] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.318479] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 574.318479] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.318479] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 574.318479] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.318479] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 574.318479] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.318479] env[62368]: ERROR nova.compute.manager raise self.value [ 574.318479] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.318479] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 574.318479] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.318479] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 574.319047] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.319047] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 574.319047] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. [ 574.319047] env[62368]: ERROR nova.compute.manager [ 574.319047] env[62368]: Traceback (most recent call last): [ 574.319047] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 574.319047] env[62368]: listener.cb(fileno) [ 574.319047] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.319047] env[62368]: result = function(*args, **kwargs) [ 574.319047] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.319047] env[62368]: return func(*args, **kwargs) [ 574.319047] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.319047] env[62368]: raise e [ 574.319047] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.319047] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 574.319047] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.319047] env[62368]: created_port_ids = self._update_ports_for_instance( [ 574.319047] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.319047] env[62368]: with excutils.save_and_reraise_exception(): [ 574.319047] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.319047] env[62368]: self.force_reraise() [ 574.319047] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.319047] env[62368]: raise self.value [ 574.319047] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.319047] env[62368]: updated_port = self._update_port( [ 574.319047] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.319047] env[62368]: _ensure_no_port_binding_failure(port) [ 574.319047] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.319047] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 574.319923] env[62368]: nova.exception.PortBindingFailed: Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. [ 574.319923] env[62368]: Removing descriptor: 15 [ 574.319923] env[62368]: ERROR nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Traceback (most recent call last): [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] yield resources [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self.driver.spawn(context, instance, image_meta, [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.319923] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] vm_ref = self.build_virtual_machine(instance, [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] for vif in network_info: [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] return self._sync_wrapper(fn, *args, **kwargs) [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self.wait() [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self[:] = self._gt.wait() [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] return self._exit_event.wait() [ 574.320337] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] result = hub.switch() [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] return self.greenlet.switch() [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] result = function(*args, **kwargs) [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] return func(*args, **kwargs) [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] raise e [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] nwinfo = self.network_api.allocate_for_instance( [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.320788] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] created_port_ids = self._update_ports_for_instance( [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] with excutils.save_and_reraise_exception(): [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self.force_reraise() [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] raise self.value [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] updated_port = self._update_port( [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] _ensure_no_port_binding_failure(port) [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.321193] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] raise exception.PortBindingFailed(port_id=port['id']) [ 574.321575] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] nova.exception.PortBindingFailed: Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. [ 574.321575] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] [ 574.321575] env[62368]: INFO nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Terminating instance [ 574.322695] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Acquiring lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.528018] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.528018] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 574.530774] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.204s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.530954] env[62368]: DEBUG nova.objects.instance [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 574.707057] env[62368]: DEBUG nova.network.neutron [req-5cbb64d8-6e1b-4474-9bff-912974d69e70 req-bcf1b4b3-c113-454d-9a47-224d343e3dc9 service nova] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.846516] env[62368]: DEBUG nova.network.neutron [req-5cbb64d8-6e1b-4474-9bff-912974d69e70 req-bcf1b4b3-c113-454d-9a47-224d343e3dc9 service nova] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.041403] env[62368]: DEBUG nova.compute.utils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.044612] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.044612] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 575.190676] env[62368]: DEBUG nova.policy [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06cf55c752484c87baff48d61a786be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee3ed873e98f4d92bedda3b94fe651f6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 575.349070] env[62368]: DEBUG oslo_concurrency.lockutils [req-5cbb64d8-6e1b-4474-9bff-912974d69e70 req-bcf1b4b3-c113-454d-9a47-224d343e3dc9 service nova] Releasing lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.349598] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Acquired lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.349786] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.544084] env[62368]: DEBUG oslo_concurrency.lockutils [None req-745c2bca-f4ea-4760-bfe8-e8a8e82da4e1 tempest-ServersAdmin275Test-1044396442 tempest-ServersAdmin275Test-1044396442-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.545247] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.126s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.552082] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 575.874977] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.892155] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Successfully created port: ef23e468-378f-4027-8c2c-94380011a2a8 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.043511] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.384307] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Acquiring lock "a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.384685] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Lock "a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.395790] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Acquiring lock "ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.396195] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Lock "ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.546560] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Releasing lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.546964] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 576.547182] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 576.550140] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bba2da16-b67c-4484-970f-40d3b7444d07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.560643] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9461e385-8760-42ab-9db1-1ed2b374b531 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.575039] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 576.587740] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 01654798-4e54-4177-b20c-f6e0aecf72a7 could not be found. [ 576.587960] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 576.588171] env[62368]: INFO nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 576.588408] env[62368]: DEBUG oslo.service.loopingcall [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.591025] env[62368]: DEBUG nova.compute.manager [-] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.591025] env[62368]: DEBUG nova.network.neutron [-] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 576.606599] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c87150-5fea-4656-a896-0b1ae74cfd8a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.610266] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 576.611274] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 576.611274] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.611274] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 576.611274] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.611274] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 576.611529] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 576.611529] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 576.611628] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 576.611783] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 576.611951] env[62368]: DEBUG nova.virt.hardware [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 576.612740] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6319196d-7fea-473c-ac48-0c4f11465e48 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.621589] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9737de93-4feb-4ed6-920a-56caf7f2deca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.625344] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76b4391-4a77-41a4-ba94-61ba37e7d79a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.631648] env[62368]: DEBUG nova.network.neutron [-] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.667683] env[62368]: DEBUG nova.network.neutron [-] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.668623] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b5129d-4abf-4bb0-a5bb-1ae622ea0e71 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.672107] env[62368]: DEBUG nova.compute.manager [req-1eeda5b7-a694-403c-83b0-f87ee25b9aa2 req-e987a343-5436-4da9-9d4f-2e5739d4758a service nova] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Received event network-vif-deleted-eade0b39-4a3a-4903-9ded-828f9f9ccc95 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.678462] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366956ef-1fe6-4375-b5d2-7901c4e1957a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.695024] env[62368]: DEBUG nova.compute.provider_tree [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.154174] env[62368]: ERROR nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. [ 577.154174] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.154174] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.154174] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.154174] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.154174] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.154174] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.154174] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.154174] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.154174] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 577.154174] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.154174] env[62368]: ERROR nova.compute.manager raise self.value [ 577.154174] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.154174] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.154174] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.154174] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.154839] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.154839] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.154839] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. [ 577.154839] env[62368]: ERROR nova.compute.manager [ 577.154839] env[62368]: Traceback (most recent call last): [ 577.154839] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.154839] env[62368]: listener.cb(fileno) [ 577.154839] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.154839] env[62368]: result = function(*args, **kwargs) [ 577.154839] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.154839] env[62368]: return func(*args, **kwargs) [ 577.154839] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.154839] env[62368]: raise e [ 577.154839] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.154839] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 577.154839] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.154839] env[62368]: created_port_ids = self._update_ports_for_instance( [ 577.154839] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.154839] env[62368]: with excutils.save_and_reraise_exception(): [ 577.154839] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.154839] env[62368]: self.force_reraise() [ 577.154839] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.154839] env[62368]: raise self.value [ 577.154839] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.154839] env[62368]: updated_port = self._update_port( [ 577.154839] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.154839] env[62368]: _ensure_no_port_binding_failure(port) [ 577.154839] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.154839] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.155663] env[62368]: nova.exception.PortBindingFailed: Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. [ 577.155663] env[62368]: Removing descriptor: 15 [ 577.155663] env[62368]: ERROR nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Traceback (most recent call last): [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] yield resources [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self.driver.spawn(context, instance, image_meta, [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.155663] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] vm_ref = self.build_virtual_machine(instance, [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] for vif in network_info: [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] return self._sync_wrapper(fn, *args, **kwargs) [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self.wait() [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self[:] = self._gt.wait() [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] return self._exit_event.wait() [ 577.156026] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] result = hub.switch() [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] return self.greenlet.switch() [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] result = function(*args, **kwargs) [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] return func(*args, **kwargs) [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] raise e [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] nwinfo = self.network_api.allocate_for_instance( [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.156418] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] created_port_ids = self._update_ports_for_instance( [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] with excutils.save_and_reraise_exception(): [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self.force_reraise() [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] raise self.value [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] updated_port = self._update_port( [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] _ensure_no_port_binding_failure(port) [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.156818] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] raise exception.PortBindingFailed(port_id=port['id']) [ 577.157186] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] nova.exception.PortBindingFailed: Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. [ 577.157186] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] [ 577.157186] env[62368]: INFO nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Terminating instance [ 577.158427] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.158656] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquired lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.158877] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 577.175088] env[62368]: INFO nova.compute.manager [-] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Took 0.58 seconds to deallocate network for instance. [ 577.178318] env[62368]: DEBUG nova.compute.claims [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 577.178504] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.195316] env[62368]: DEBUG nova.scheduler.client.report [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.702416] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.155s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.702416] env[62368]: ERROR nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. [ 577.702416] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Traceback (most recent call last): [ 577.702416] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.702416] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self.driver.spawn(context, instance, image_meta, [ 577.702416] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.702416] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.702416] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.702416] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] vm_ref = self.build_virtual_machine(instance, [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] for vif in network_info: [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] return self._sync_wrapper(fn, *args, **kwargs) [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self.wait() [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self[:] = self._gt.wait() [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] return self._exit_event.wait() [ 577.702806] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] result = hub.switch() [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] return self.greenlet.switch() [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] result = function(*args, **kwargs) [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] return func(*args, **kwargs) [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] raise e [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] nwinfo = self.network_api.allocate_for_instance( [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.703208] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] created_port_ids = self._update_ports_for_instance( [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] with excutils.save_and_reraise_exception(): [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] self.force_reraise() [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] raise self.value [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] updated_port = self._update_port( [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] _ensure_no_port_binding_failure(port) [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.703633] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] raise exception.PortBindingFailed(port_id=port['id']) [ 577.704205] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] nova.exception.PortBindingFailed: Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. [ 577.704205] env[62368]: ERROR nova.compute.manager [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] [ 577.704205] env[62368]: DEBUG nova.compute.utils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.704894] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.101s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.705137] env[62368]: DEBUG nova.objects.instance [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lazy-loading 'resources' on Instance uuid 0eb2a772-7635-481a-ae2c-9f23669dd175 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 577.706757] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Build of instance 54e3c531-9253-4fc0-999c-2cb9a8388d08 was re-scheduled: Binding failed for port e2385327-f738-43e8-bfbc-6503b85ad45e, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.707226] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.707502] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquiring lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.707591] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Acquired lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.707752] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 577.756330] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.853269] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.234117] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.351324] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.355778] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Releasing lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.356253] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 578.356436] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 578.358340] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63c76fa7-1b3f-4446-bc94-511f20a74413 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.374362] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e5895a-7163-4d21-beec-6efb0713bd28 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.397519] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9fe2763-2d3c-48f6-a836-e0c8634acc95 could not be found. [ 578.397747] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 578.397925] env[62368]: INFO nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Took 0.04 seconds to destroy the instance on the hypervisor. [ 578.398234] env[62368]: DEBUG oslo.service.loopingcall [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.400678] env[62368]: DEBUG nova.compute.manager [-] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.400784] env[62368]: DEBUG nova.network.neutron [-] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.420993] env[62368]: DEBUG nova.network.neutron [-] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.646099] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833aed48-7a9c-467f-9f21-b2e0d04a1837 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.654719] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70f29b1-3835-4c16-aad0-af3033812ac1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.684771] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72de9998-3d22-4a65-90a0-e27eea0b1ad1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.691067] env[62368]: DEBUG nova.compute.manager [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Received event network-changed-ef23e468-378f-4027-8c2c-94380011a2a8 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 578.691241] env[62368]: DEBUG nova.compute.manager [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Refreshing instance network info cache due to event network-changed-ef23e468-378f-4027-8c2c-94380011a2a8. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 578.691360] env[62368]: DEBUG oslo_concurrency.lockutils [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] Acquiring lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.691523] env[62368]: DEBUG oslo_concurrency.lockutils [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] Acquired lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.691686] env[62368]: DEBUG nova.network.neutron [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Refreshing network info cache for port ef23e468-378f-4027-8c2c-94380011a2a8 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 578.698114] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a8c596-4e30-44b6-8a2a-5fbdc07c131e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.714467] env[62368]: DEBUG nova.compute.provider_tree [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.858144] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Releasing lock "refresh_cache-54e3c531-9253-4fc0-999c-2cb9a8388d08" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.858396] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 578.858575] env[62368]: DEBUG nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.858738] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.888724] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.924556] env[62368]: DEBUG nova.network.neutron [-] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.217715] env[62368]: DEBUG nova.network.neutron [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.221587] env[62368]: DEBUG nova.scheduler.client.report [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.283560] env[62368]: DEBUG nova.network.neutron [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.394409] env[62368]: DEBUG nova.network.neutron [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.427235] env[62368]: INFO nova.compute.manager [-] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Took 1.03 seconds to deallocate network for instance. [ 579.429649] env[62368]: DEBUG nova.compute.claims [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 579.429822] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.730491] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.023s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.736365] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.465s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.736679] env[62368]: INFO nova.compute.claims [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.764435] env[62368]: INFO nova.scheduler.client.report [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Deleted allocations for instance 0eb2a772-7635-481a-ae2c-9f23669dd175 [ 579.786844] env[62368]: DEBUG oslo_concurrency.lockutils [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] Releasing lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.787156] env[62368]: DEBUG nova.compute.manager [req-bc0a7379-c308-40a6-bca1-aad41ad0e146 req-4a5ede9b-d351-451f-b64b-0a723549c3a8 service nova] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Received event network-vif-deleted-ef23e468-378f-4027-8c2c-94380011a2a8 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.897232] env[62368]: INFO nova.compute.manager [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] [instance: 54e3c531-9253-4fc0-999c-2cb9a8388d08] Took 1.04 seconds to deallocate network for instance. [ 580.276100] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3813257b-7542-4e39-a4b2-97279cd20e1e tempest-ServersAdmin275Test-241181603 tempest-ServersAdmin275Test-241181603-project-member] Lock "0eb2a772-7635-481a-ae2c-9f23669dd175" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.682s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.954154] env[62368]: INFO nova.scheduler.client.report [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Deleted allocations for instance 54e3c531-9253-4fc0-999c-2cb9a8388d08 [ 581.272782] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48dadcd5-346a-40d5-9f0d-251d8648f8c9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.282681] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6501c7d4-3f46-45d1-b2f8-a916c2fc30b2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.322208] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82410b8-23f0-42a0-86c0-c3b3a7579b8e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.327276] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b5b9c2-723e-4b09-a2e9-ce23a434daa2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.342477] env[62368]: DEBUG nova.compute.provider_tree [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.463540] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1bd64ab3-5e7c-459f-941b-6fd879bc82dc tempest-ListServerFiltersTestJSON-67153098 tempest-ListServerFiltersTestJSON-67153098-project-member] Lock "54e3c531-9253-4fc0-999c-2cb9a8388d08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.839s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.848116] env[62368]: DEBUG nova.scheduler.client.report [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.967574] env[62368]: DEBUG nova.compute.manager [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.353292] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.353815] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.358213] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.033s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.358565] env[62368]: INFO nova.compute.claims [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.491711] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.863704] env[62368]: DEBUG nova.compute.utils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.869379] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.869688] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 582.943264] env[62368]: DEBUG nova.policy [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed164cbe56ba4481bd8fa750836466be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '016e17c9deea4318a0e4b09ac9adf4d7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.057978] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquiring lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.058239] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.372476] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.398088] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Successfully created port: ee270188-766a-470c-bc9d-ac8f401df526 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.866416] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5787cd-3314-461f-8399-d57ed9e4b25d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.874633] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf1a74f-631c-4eae-8cba-f4247ebeda69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.914405] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29573703-c4bc-4f40-a800-d0053c8eb2aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.924854] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df9df88-0ee4-464e-87c7-6e1d39d4aab2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.944163] env[62368]: DEBUG nova.compute.provider_tree [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.383713] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.416617] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:12:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='63616242',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-905768029',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.416883] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.417053] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.417238] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.417383] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.417527] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.419501] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.419501] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.419501] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.419501] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.419723] env[62368]: DEBUG nova.virt.hardware [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.420580] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc126eb-2313-45d0-b996-196846fee65d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.432525] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524843ea-a278-46a0-b216-8efbd33d9ecd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.450592] env[62368]: DEBUG nova.scheduler.client.report [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.586883] env[62368]: DEBUG nova.compute.manager [req-9125f193-8f85-4a1d-95fe-43b09af9d01c req-ea5f0f67-2ca5-422b-be14-85eb1bf61ca8 service nova] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Received event network-changed-ee270188-766a-470c-bc9d-ac8f401df526 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.587059] env[62368]: DEBUG nova.compute.manager [req-9125f193-8f85-4a1d-95fe-43b09af9d01c req-ea5f0f67-2ca5-422b-be14-85eb1bf61ca8 service nova] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Refreshing instance network info cache due to event network-changed-ee270188-766a-470c-bc9d-ac8f401df526. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 584.587156] env[62368]: DEBUG oslo_concurrency.lockutils [req-9125f193-8f85-4a1d-95fe-43b09af9d01c req-ea5f0f67-2ca5-422b-be14-85eb1bf61ca8 service nova] Acquiring lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.587300] env[62368]: DEBUG oslo_concurrency.lockutils [req-9125f193-8f85-4a1d-95fe-43b09af9d01c req-ea5f0f67-2ca5-422b-be14-85eb1bf61ca8 service nova] Acquired lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.588419] env[62368]: DEBUG nova.network.neutron [req-9125f193-8f85-4a1d-95fe-43b09af9d01c req-ea5f0f67-2ca5-422b-be14-85eb1bf61ca8 service nova] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Refreshing network info cache for port ee270188-766a-470c-bc9d-ac8f401df526 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 584.698214] env[62368]: ERROR nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. [ 584.698214] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.698214] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.698214] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.698214] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.698214] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.698214] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.698214] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.698214] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.698214] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 584.698214] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.698214] env[62368]: ERROR nova.compute.manager raise self.value [ 584.698214] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.698214] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.698214] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.698214] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.698834] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.698834] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.698834] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. [ 584.698834] env[62368]: ERROR nova.compute.manager [ 584.698834] env[62368]: Traceback (most recent call last): [ 584.698834] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.698834] env[62368]: listener.cb(fileno) [ 584.698834] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.698834] env[62368]: result = function(*args, **kwargs) [ 584.698834] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.698834] env[62368]: return func(*args, **kwargs) [ 584.698834] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.698834] env[62368]: raise e [ 584.698834] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.698834] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 584.698834] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.698834] env[62368]: created_port_ids = self._update_ports_for_instance( [ 584.698834] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.698834] env[62368]: with excutils.save_and_reraise_exception(): [ 584.698834] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.698834] env[62368]: self.force_reraise() [ 584.698834] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.698834] env[62368]: raise self.value [ 584.698834] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.698834] env[62368]: updated_port = self._update_port( [ 584.698834] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.698834] env[62368]: _ensure_no_port_binding_failure(port) [ 584.698834] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.698834] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.699846] env[62368]: nova.exception.PortBindingFailed: Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. [ 584.699846] env[62368]: Removing descriptor: 15 [ 584.699846] env[62368]: ERROR nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Traceback (most recent call last): [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] yield resources [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self.driver.spawn(context, instance, image_meta, [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.699846] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] vm_ref = self.build_virtual_machine(instance, [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] for vif in network_info: [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] return self._sync_wrapper(fn, *args, **kwargs) [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self.wait() [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self[:] = self._gt.wait() [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] return self._exit_event.wait() [ 584.700300] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] result = hub.switch() [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] return self.greenlet.switch() [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] result = function(*args, **kwargs) [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] return func(*args, **kwargs) [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] raise e [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] nwinfo = self.network_api.allocate_for_instance( [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.700763] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] created_port_ids = self._update_ports_for_instance( [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] with excutils.save_and_reraise_exception(): [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self.force_reraise() [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] raise self.value [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] updated_port = self._update_port( [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] _ensure_no_port_binding_failure(port) [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.701219] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] raise exception.PortBindingFailed(port_id=port['id']) [ 584.701645] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] nova.exception.PortBindingFailed: Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. [ 584.701645] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] [ 584.701645] env[62368]: INFO nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Terminating instance [ 584.703027] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Acquiring lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.957517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.958092] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 584.961642] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.552s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.963397] env[62368]: INFO nova.compute.claims [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.103965] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquiring lock "a5c638d8-f761-497d-bde2-39ea90d8d151" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.104225] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "a5c638d8-f761-497d-bde2-39ea90d8d151" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.124619] env[62368]: DEBUG nova.network.neutron [req-9125f193-8f85-4a1d-95fe-43b09af9d01c req-ea5f0f67-2ca5-422b-be14-85eb1bf61ca8 service nova] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.278360] env[62368]: DEBUG nova.network.neutron [req-9125f193-8f85-4a1d-95fe-43b09af9d01c req-ea5f0f67-2ca5-422b-be14-85eb1bf61ca8 service nova] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.468517] env[62368]: DEBUG nova.compute.utils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.472852] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.472941] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 585.551636] env[62368]: DEBUG nova.policy [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9c7565b4dd942028a9d936cb71b0d91', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec576ba132b946b2a72e8c18282c3914', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.781047] env[62368]: DEBUG oslo_concurrency.lockutils [req-9125f193-8f85-4a1d-95fe-43b09af9d01c req-ea5f0f67-2ca5-422b-be14-85eb1bf61ca8 service nova] Releasing lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.781486] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Acquired lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.781959] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.977021] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.313541] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.338646] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Successfully created port: 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.446596] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.487668] env[62368]: INFO nova.virt.block_device [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Booting with volume 089c08f9-2c95-4a35-885a-31ffba68be97 at /dev/sda [ 586.495749] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b44ccdd-ab51-412e-886b-4fcfeb49a8a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.515712] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a12e99-39e9-4e20-b905-5ab612754b82 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.547408] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3ee449-898d-4d23-80bd-9f5aa07d55d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.554887] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb88d47-2443-4858-aef6-1d4230c0994d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.559566] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac247418-1587-43f4-8e72-1d00299eb1b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.570977] env[62368]: DEBUG nova.compute.provider_tree [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.575962] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19ad8e6-7661-432e-9b0e-e8a05fbc24bb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.587237] env[62368]: DEBUG nova.scheduler.client.report [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.603598] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4920910-4e39-4f1c-8d37-03ff5ced6962 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.611940] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633b2cc1-e3eb-48e5-8c3d-2c513c8e7b35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.624185] env[62368]: DEBUG nova.compute.manager [req-bafc3840-6746-4880-92eb-60264fc87d8b req-02199d47-0c21-4610-b444-44ac8410d3e7 service nova] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Received event network-vif-deleted-ee270188-766a-470c-bc9d-ac8f401df526 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.635980] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9ed7e6-cc4e-4d16-8334-252a4fb5da5b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.642108] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cac9148-b80a-4e62-ba57-5daaa0d49be2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.656199] env[62368]: DEBUG nova.virt.block_device [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Updating existing volume attachment record: ed439d34-1945-4c98-b026-01538da9364e {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 586.957349] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Releasing lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.958214] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.958214] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 586.958368] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d21a6d9d-4529-417c-b289-6b18e6f21342 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.970900] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159010e7-147a-420b-a905-c552abfb9302 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.992357] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23 could not be found. [ 586.992624] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 586.995362] env[62368]: INFO nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Took 0.03 seconds to destroy the instance on the hypervisor. [ 586.995362] env[62368]: DEBUG oslo.service.loopingcall [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.995362] env[62368]: DEBUG nova.compute.manager [-] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.995362] env[62368]: DEBUG nova.network.neutron [-] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.029759] env[62368]: DEBUG nova.network.neutron [-] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.091807] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.130s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.092221] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.095515] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.483s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.097063] env[62368]: INFO nova.compute.claims [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.516831] env[62368]: DEBUG nova.network.neutron [-] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.603933] env[62368]: DEBUG nova.compute.utils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.606819] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 587.606996] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 587.670955] env[62368]: DEBUG nova.policy [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06cf55c752484c87baff48d61a786be0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee3ed873e98f4d92bedda3b94fe651f6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.985817] env[62368]: ERROR nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. [ 587.985817] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.985817] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.985817] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.985817] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.985817] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.985817] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.985817] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.985817] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.985817] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 587.985817] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.985817] env[62368]: ERROR nova.compute.manager raise self.value [ 587.985817] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.985817] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.985817] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.985817] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.986780] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.986780] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.986780] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. [ 587.986780] env[62368]: ERROR nova.compute.manager [ 587.986780] env[62368]: Traceback (most recent call last): [ 587.986780] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.986780] env[62368]: listener.cb(fileno) [ 587.986780] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.986780] env[62368]: result = function(*args, **kwargs) [ 587.986780] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.986780] env[62368]: return func(*args, **kwargs) [ 587.986780] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.986780] env[62368]: raise e [ 587.986780] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.986780] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 587.986780] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.986780] env[62368]: created_port_ids = self._update_ports_for_instance( [ 587.986780] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.986780] env[62368]: with excutils.save_and_reraise_exception(): [ 587.986780] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.986780] env[62368]: self.force_reraise() [ 587.986780] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.986780] env[62368]: raise self.value [ 587.986780] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.986780] env[62368]: updated_port = self._update_port( [ 587.986780] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.986780] env[62368]: _ensure_no_port_binding_failure(port) [ 587.986780] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.986780] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.987353] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. [ 587.987353] env[62368]: Removing descriptor: 15 [ 588.022105] env[62368]: INFO nova.compute.manager [-] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Took 1.03 seconds to deallocate network for instance. [ 588.024030] env[62368]: DEBUG nova.compute.claims [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.024321] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.037838] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Successfully created port: bc51ee6a-376f-4ffc-be51-e6e61d8acca9 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.107434] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 588.605891] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6646847e-a24c-4e93-b58d-d099bcfac14a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.613905] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c5f668-e0c3-423a-bcf8-0bcd0716360d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.655020] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f765596-7784-4ae4-91a0-f2338cba43d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.663315] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746aa0cd-bb33-45a8-b394-496b9442ebdb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.678754] env[62368]: DEBUG nova.compute.provider_tree [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.695380] env[62368]: DEBUG nova.compute.manager [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Received event network-changed-80af8aeb-f5f2-48a3-a218-8f0fab2efc6a {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.695563] env[62368]: DEBUG nova.compute.manager [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Refreshing instance network info cache due to event network-changed-80af8aeb-f5f2-48a3-a218-8f0fab2efc6a. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 588.695784] env[62368]: DEBUG oslo_concurrency.lockutils [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] Acquiring lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.695916] env[62368]: DEBUG oslo_concurrency.lockutils [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] Acquired lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.696086] env[62368]: DEBUG nova.network.neutron [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Refreshing network info cache for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 588.788374] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 588.789984] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 588.789984] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 588.790129] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.790236] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 588.790379] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.790520] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 588.790718] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 588.790871] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 588.791043] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 588.791207] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 588.791381] env[62368]: DEBUG nova.virt.hardware [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 588.792320] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47682ca0-3dcd-42ae-b8c8-564ce7097228 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.801238] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e20b228-6a4e-4a00-af23-43fd58021ffe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.816042] env[62368]: ERROR nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Traceback (most recent call last): [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] yield resources [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self.driver.spawn(context, instance, image_meta, [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] vm_ref = self.build_virtual_machine(instance, [ 588.816042] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] for vif in network_info: [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] return self._sync_wrapper(fn, *args, **kwargs) [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self.wait() [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self[:] = self._gt.wait() [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] return self._exit_event.wait() [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 588.816414] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] current.throw(*self._exc) [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] result = function(*args, **kwargs) [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] return func(*args, **kwargs) [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] raise e [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] nwinfo = self.network_api.allocate_for_instance( [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] created_port_ids = self._update_ports_for_instance( [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] with excutils.save_and_reraise_exception(): [ 588.816666] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self.force_reraise() [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] raise self.value [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] updated_port = self._update_port( [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] _ensure_no_port_binding_failure(port) [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] raise exception.PortBindingFailed(port_id=port['id']) [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] nova.exception.PortBindingFailed: Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. [ 588.816930] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] [ 588.816930] env[62368]: INFO nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Terminating instance [ 588.818668] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Acquiring lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.157098] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.181349] env[62368]: DEBUG nova.scheduler.client.report [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.187658] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.187879] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.188038] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.188232] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.188389] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.188537] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.188738] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.188899] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.189066] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.189296] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.189945] env[62368]: DEBUG nova.virt.hardware [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.190270] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c915c0d2-d9ed-4090-a65c-13d9b6b0f61b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.201154] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33daea8e-9159-4f78-a73b-b87b08576cfb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.229153] env[62368]: DEBUG nova.network.neutron [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.354344] env[62368]: DEBUG nova.network.neutron [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.512018] env[62368]: DEBUG nova.compute.manager [req-dea5425b-27ae-4781-81d2-dfbe61506962 req-17a69847-1785-466c-90c8-7ae78293dfe7 service nova] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Received event network-changed-bc51ee6a-376f-4ffc-be51-e6e61d8acca9 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.512018] env[62368]: DEBUG nova.compute.manager [req-dea5425b-27ae-4781-81d2-dfbe61506962 req-17a69847-1785-466c-90c8-7ae78293dfe7 service nova] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Refreshing instance network info cache due to event network-changed-bc51ee6a-376f-4ffc-be51-e6e61d8acca9. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 589.512018] env[62368]: DEBUG oslo_concurrency.lockutils [req-dea5425b-27ae-4781-81d2-dfbe61506962 req-17a69847-1785-466c-90c8-7ae78293dfe7 service nova] Acquiring lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.512018] env[62368]: DEBUG oslo_concurrency.lockutils [req-dea5425b-27ae-4781-81d2-dfbe61506962 req-17a69847-1785-466c-90c8-7ae78293dfe7 service nova] Acquired lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.512018] env[62368]: DEBUG nova.network.neutron [req-dea5425b-27ae-4781-81d2-dfbe61506962 req-17a69847-1785-466c-90c8-7ae78293dfe7 service nova] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Refreshing network info cache for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 589.654022] env[62368]: ERROR nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. [ 589.654022] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.654022] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.654022] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.654022] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.654022] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.654022] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.654022] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.654022] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.654022] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 589.654022] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.654022] env[62368]: ERROR nova.compute.manager raise self.value [ 589.654022] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.654022] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.654022] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.654022] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.654561] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.654561] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.654561] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. [ 589.654561] env[62368]: ERROR nova.compute.manager [ 589.654561] env[62368]: Traceback (most recent call last): [ 589.654561] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.654561] env[62368]: listener.cb(fileno) [ 589.654561] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.654561] env[62368]: result = function(*args, **kwargs) [ 589.654561] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.654561] env[62368]: return func(*args, **kwargs) [ 589.654561] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.654561] env[62368]: raise e [ 589.654561] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.654561] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 589.654561] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.654561] env[62368]: created_port_ids = self._update_ports_for_instance( [ 589.654561] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.654561] env[62368]: with excutils.save_and_reraise_exception(): [ 589.654561] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.654561] env[62368]: self.force_reraise() [ 589.654561] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.654561] env[62368]: raise self.value [ 589.654561] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.654561] env[62368]: updated_port = self._update_port( [ 589.654561] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.654561] env[62368]: _ensure_no_port_binding_failure(port) [ 589.654561] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.654561] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.655287] env[62368]: nova.exception.PortBindingFailed: Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. [ 589.655287] env[62368]: Removing descriptor: 18 [ 589.655287] env[62368]: ERROR nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Traceback (most recent call last): [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] yield resources [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self.driver.spawn(context, instance, image_meta, [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.655287] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] vm_ref = self.build_virtual_machine(instance, [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] for vif in network_info: [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] return self._sync_wrapper(fn, *args, **kwargs) [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self.wait() [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self[:] = self._gt.wait() [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] return self._exit_event.wait() [ 589.655544] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] result = hub.switch() [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] return self.greenlet.switch() [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] result = function(*args, **kwargs) [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] return func(*args, **kwargs) [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] raise e [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] nwinfo = self.network_api.allocate_for_instance( [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.655798] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] created_port_ids = self._update_ports_for_instance( [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] with excutils.save_and_reraise_exception(): [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self.force_reraise() [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] raise self.value [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] updated_port = self._update_port( [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] _ensure_no_port_binding_failure(port) [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.656058] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] raise exception.PortBindingFailed(port_id=port['id']) [ 589.656296] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] nova.exception.PortBindingFailed: Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. [ 589.656296] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] [ 589.656296] env[62368]: INFO nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Terminating instance [ 589.658353] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.695049] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.695465] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.698099] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.616s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.702017] env[62368]: INFO nova.compute.claims [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.856994] env[62368]: DEBUG oslo_concurrency.lockutils [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] Releasing lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.857296] env[62368]: DEBUG nova.compute.manager [req-5f5ef26e-96ed-4f70-aade-72b7fef1044e req-973cc994-dbc5-4384-9522-a12d84ed7192 service nova] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Received event network-vif-deleted-80af8aeb-f5f2-48a3-a218-8f0fab2efc6a {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.857628] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Acquired lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.857874] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 590.027364] env[62368]: DEBUG nova.network.neutron [req-dea5425b-27ae-4781-81d2-dfbe61506962 req-17a69847-1785-466c-90c8-7ae78293dfe7 service nova] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.097130] env[62368]: DEBUG nova.network.neutron [req-dea5425b-27ae-4781-81d2-dfbe61506962 req-17a69847-1785-466c-90c8-7ae78293dfe7 service nova] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.107037] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.107266] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.204208] env[62368]: DEBUG nova.compute.utils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.209070] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 590.209070] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 590.252452] env[62368]: DEBUG nova.policy [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '83b960e3cb1747cca6b5d30b362d7c5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b4bbe6b942c40d3a046937ef8367370', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.375540] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.445931] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.522191] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Successfully created port: 38d0cf3e-d303-4caf-8bc9-0895237d59a1 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.599913] env[62368]: DEBUG oslo_concurrency.lockutils [req-dea5425b-27ae-4781-81d2-dfbe61506962 req-17a69847-1785-466c-90c8-7ae78293dfe7 service nova] Releasing lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.600345] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquired lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.600530] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 590.713021] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.953209] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Releasing lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.953209] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.953209] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d441229e-ea8b-4a60-87b1-0d03ea769632 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.967291] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5664fb-0a18-4577-9d37-17253c9ed792 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.991839] env[62368]: WARNING nova.virt.vmwareapi.driver [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 546a40aa-c278-460c-8c91-c6c903d751ab could not be found. [ 590.992091] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 590.994854] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d01dc15f-af54-4b29-b569-9a2c7335e9fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.002951] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80903458-8f5d-409c-98be-d7cbdf3a156c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.027106] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 546a40aa-c278-460c-8c91-c6c903d751ab could not be found. [ 591.027324] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 591.027506] env[62368]: INFO nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Took 0.07 seconds to destroy the instance on the hypervisor. [ 591.027817] env[62368]: DEBUG oslo.service.loopingcall [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.030403] env[62368]: DEBUG nova.compute.manager [-] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.030506] env[62368]: DEBUG nova.network.neutron [-] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 591.048362] env[62368]: DEBUG nova.network.neutron [-] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.121458] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.170203] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e79755-9add-4727-bbe7-9b8ff6e39908 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.177949] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e40e65-7257-4dac-8db6-fcb72a7b15d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.217007] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.218643] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183daeda-c9f1-4f33-841a-6ef9a1b13699 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.229417] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a0e8c2-2fe5-44bd-8c0a-9cea0e6cb0fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.246153] env[62368]: DEBUG nova.compute.provider_tree [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.546441] env[62368]: ERROR nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. [ 591.546441] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.546441] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.546441] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.546441] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.546441] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.546441] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.546441] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.546441] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.546441] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 591.546441] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.546441] env[62368]: ERROR nova.compute.manager raise self.value [ 591.546441] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.546441] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.546441] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.546441] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.546823] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.546823] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.546823] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. [ 591.546823] env[62368]: ERROR nova.compute.manager [ 591.546823] env[62368]: Traceback (most recent call last): [ 591.546823] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.546823] env[62368]: listener.cb(fileno) [ 591.546823] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.546823] env[62368]: result = function(*args, **kwargs) [ 591.546823] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.546823] env[62368]: return func(*args, **kwargs) [ 591.546823] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.546823] env[62368]: raise e [ 591.546823] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.546823] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 591.546823] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.546823] env[62368]: created_port_ids = self._update_ports_for_instance( [ 591.546823] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.546823] env[62368]: with excutils.save_and_reraise_exception(): [ 591.546823] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.546823] env[62368]: self.force_reraise() [ 591.546823] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.546823] env[62368]: raise self.value [ 591.546823] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.546823] env[62368]: updated_port = self._update_port( [ 591.546823] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.546823] env[62368]: _ensure_no_port_binding_failure(port) [ 591.546823] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.546823] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.547501] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. [ 591.547501] env[62368]: Removing descriptor: 18 [ 591.548900] env[62368]: DEBUG nova.compute.manager [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Received event network-vif-deleted-bc51ee6a-376f-4ffc-be51-e6e61d8acca9 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.549109] env[62368]: DEBUG nova.compute.manager [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Received event network-changed-38d0cf3e-d303-4caf-8bc9-0895237d59a1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.549272] env[62368]: DEBUG nova.compute.manager [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Refreshing instance network info cache due to event network-changed-38d0cf3e-d303-4caf-8bc9-0895237d59a1. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 591.549480] env[62368]: DEBUG oslo_concurrency.lockutils [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] Acquiring lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.549613] env[62368]: DEBUG oslo_concurrency.lockutils [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] Acquired lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.549769] env[62368]: DEBUG nova.network.neutron [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Refreshing network info cache for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 591.551244] env[62368]: DEBUG nova.network.neutron [-] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.726112] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Releasing lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.726608] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 591.726803] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 591.727953] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.730355] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b91c9cfd-1749-4989-a6bd-3e2d052dd991 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.739721] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed2c07d-eb3b-4b9b-88cf-d4e7c61268e1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.751741] env[62368]: DEBUG nova.scheduler.client.report [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.762245] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.762500] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.762661] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.762855] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.762999] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.763178] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.763436] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.763584] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.763750] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.763910] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.764095] env[62368]: DEBUG nova.virt.hardware [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.764943] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d92d9a-bbfa-4255-97bd-71bb78e615a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.772647] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fd5478d6-9fa8-4471-8b95-b89f3d826bcc could not be found. [ 591.772770] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 591.772940] env[62368]: INFO nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 591.773197] env[62368]: DEBUG oslo.service.loopingcall [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.773762] env[62368]: DEBUG nova.compute.manager [-] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.773862] env[62368]: DEBUG nova.network.neutron [-] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 591.779323] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad77f147-fb29-4001-be03-16984bdd5d34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.795262] env[62368]: ERROR nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Traceback (most recent call last): [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] yield resources [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self.driver.spawn(context, instance, image_meta, [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] vm_ref = self.build_virtual_machine(instance, [ 591.795262] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] for vif in network_info: [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] return self._sync_wrapper(fn, *args, **kwargs) [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self.wait() [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self[:] = self._gt.wait() [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] return self._exit_event.wait() [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 591.795659] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] current.throw(*self._exc) [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] result = function(*args, **kwargs) [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] return func(*args, **kwargs) [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] raise e [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] nwinfo = self.network_api.allocate_for_instance( [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] created_port_ids = self._update_ports_for_instance( [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] with excutils.save_and_reraise_exception(): [ 591.796093] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self.force_reraise() [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] raise self.value [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] updated_port = self._update_port( [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] _ensure_no_port_binding_failure(port) [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] raise exception.PortBindingFailed(port_id=port['id']) [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] nova.exception.PortBindingFailed: Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. [ 591.796440] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] [ 591.796440] env[62368]: INFO nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Terminating instance [ 591.797549] env[62368]: DEBUG nova.network.neutron [-] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.802039] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.054979] env[62368]: INFO nova.compute.manager [-] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Took 1.02 seconds to deallocate network for instance. [ 592.071814] env[62368]: DEBUG nova.network.neutron [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.165294] env[62368]: DEBUG nova.network.neutron [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.257236] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.257768] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 592.260307] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.575s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.299673] env[62368]: DEBUG nova.network.neutron [-] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.613186] env[62368]: INFO nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Took 0.56 seconds to detach 1 volumes for instance. [ 592.615571] env[62368]: DEBUG nova.compute.claims [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.615789] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.667644] env[62368]: DEBUG oslo_concurrency.lockutils [req-07391581-094b-4d78-9713-8bd1cae88729 req-f7714b9c-3963-4f9e-a7b6-a05d62449836 service nova] Releasing lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.668069] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquired lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.668250] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.764938] env[62368]: DEBUG nova.compute.utils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.769598] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 592.769788] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 592.801467] env[62368]: INFO nova.compute.manager [-] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Took 1.03 seconds to deallocate network for instance. [ 592.803760] env[62368]: DEBUG nova.compute.claims [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.803934] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.830467] env[62368]: DEBUG nova.policy [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0395c0d4a4a43c69ab02118d644fc6c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca340d822ef649c08b3f121df941809d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 593.174777] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb6c758-02a8-4e5f-8e51-0fdb7d359264 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.183601] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690149d3-7e4e-4dd3-9284-565f1d7682d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.222128] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.224478] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35e960e-d8b8-4377-bf92-77c103ae5702 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.232435] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f726c3ec-d896-4062-87fa-3ed6db009d9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.239026] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Successfully created port: efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.251786] env[62368]: DEBUG nova.compute.provider_tree [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.270244] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 593.341166] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.584618] env[62368]: DEBUG nova.compute.manager [req-3beb4ca8-c980-4fcb-a905-01648eaf0630 req-aa6cbcf0-9dd5-43e6-b60e-001ab5b4a07d service nova] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Received event network-vif-deleted-38d0cf3e-d303-4caf-8bc9-0895237d59a1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.757585] env[62368]: DEBUG nova.scheduler.client.report [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.843914] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Releasing lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.844350] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.844542] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 593.845149] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e5551de-2cf8-4cd6-a343-b95a5fe6ebb1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.854576] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b9726f-2287-497d-9aca-027e58d08409 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.875817] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fc4267ff-dda1-444f-a5b0-a4954626b42d could not be found. [ 593.875817] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 593.875933] env[62368]: INFO nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 593.876203] env[62368]: DEBUG oslo.service.loopingcall [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.876406] env[62368]: DEBUG nova.compute.manager [-] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.876494] env[62368]: DEBUG nova.network.neutron [-] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 593.894185] env[62368]: DEBUG nova.network.neutron [-] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.247373] env[62368]: ERROR nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. [ 594.247373] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.247373] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.247373] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.247373] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.247373] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.247373] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.247373] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.247373] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.247373] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 594.247373] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.247373] env[62368]: ERROR nova.compute.manager raise self.value [ 594.247373] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.247373] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.247373] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.247373] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.247836] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.247836] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.247836] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. [ 594.247836] env[62368]: ERROR nova.compute.manager [ 594.247836] env[62368]: Traceback (most recent call last): [ 594.247836] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.247836] env[62368]: listener.cb(fileno) [ 594.247836] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.247836] env[62368]: result = function(*args, **kwargs) [ 594.247836] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.247836] env[62368]: return func(*args, **kwargs) [ 594.247836] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.247836] env[62368]: raise e [ 594.247836] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.247836] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 594.247836] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.247836] env[62368]: created_port_ids = self._update_ports_for_instance( [ 594.247836] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.247836] env[62368]: with excutils.save_and_reraise_exception(): [ 594.247836] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.247836] env[62368]: self.force_reraise() [ 594.247836] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.247836] env[62368]: raise self.value [ 594.247836] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.247836] env[62368]: updated_port = self._update_port( [ 594.247836] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.247836] env[62368]: _ensure_no_port_binding_failure(port) [ 594.247836] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.247836] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.248571] env[62368]: nova.exception.PortBindingFailed: Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. [ 594.248571] env[62368]: Removing descriptor: 18 [ 594.262867] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.002s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.263275] env[62368]: ERROR nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Traceback (most recent call last): [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self.driver.spawn(context, instance, image_meta, [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] vm_ref = self.build_virtual_machine(instance, [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.263275] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] for vif in network_info: [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] return self._sync_wrapper(fn, *args, **kwargs) [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self.wait() [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self[:] = self._gt.wait() [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] return self._exit_event.wait() [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] result = hub.switch() [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.263575] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] return self.greenlet.switch() [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] result = function(*args, **kwargs) [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] return func(*args, **kwargs) [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] raise e [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] nwinfo = self.network_api.allocate_for_instance( [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] created_port_ids = self._update_ports_for_instance( [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] with excutils.save_and_reraise_exception(): [ 594.263946] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] self.force_reraise() [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] raise self.value [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] updated_port = self._update_port( [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] _ensure_no_port_binding_failure(port) [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] raise exception.PortBindingFailed(port_id=port['id']) [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] nova.exception.PortBindingFailed: Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. [ 594.264259] env[62368]: ERROR nova.compute.manager [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] [ 594.264539] env[62368]: DEBUG nova.compute.utils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 594.265757] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Build of instance eac92d1f-d160-4e06-a71c-d064acdffb19 was re-scheduled: Binding failed for port 59966839-e7d4-4c37-a1ba-018a59b8d38e, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 594.266218] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 594.266449] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Acquiring lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.266598] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Acquired lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.266759] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 594.267792] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.098s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.269729] env[62368]: INFO nova.compute.claims [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.279641] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.305255] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.305495] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.305727] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.305921] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.306073] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.306218] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.306419] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.306570] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.306730] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.306891] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.307067] env[62368]: DEBUG nova.virt.hardware [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.307908] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ca1ad4-1cf6-4a2f-9bd7-08b8bb2e902f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.316759] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfb4318-1ad7-419a-b6c0-4bca74bea938 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.331922] env[62368]: ERROR nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Traceback (most recent call last): [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] yield resources [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self.driver.spawn(context, instance, image_meta, [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] vm_ref = self.build_virtual_machine(instance, [ 594.331922] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] for vif in network_info: [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] return self._sync_wrapper(fn, *args, **kwargs) [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self.wait() [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self[:] = self._gt.wait() [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] return self._exit_event.wait() [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 594.332306] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] current.throw(*self._exc) [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] result = function(*args, **kwargs) [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] return func(*args, **kwargs) [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] raise e [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] nwinfo = self.network_api.allocate_for_instance( [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] created_port_ids = self._update_ports_for_instance( [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] with excutils.save_and_reraise_exception(): [ 594.332590] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self.force_reraise() [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] raise self.value [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] updated_port = self._update_port( [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] _ensure_no_port_binding_failure(port) [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] raise exception.PortBindingFailed(port_id=port['id']) [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] nova.exception.PortBindingFailed: Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. [ 594.332857] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] [ 594.332857] env[62368]: INFO nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Terminating instance [ 594.334111] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Acquiring lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.334269] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Acquired lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.334431] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 594.397056] env[62368]: DEBUG nova.network.neutron [-] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.792751] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.852240] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.885207] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.894797] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.898856] env[62368]: INFO nova.compute.manager [-] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Took 1.02 seconds to deallocate network for instance. [ 594.900915] env[62368]: DEBUG nova.compute.claims [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.901049] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.387303] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Releasing lock "refresh_cache-eac92d1f-d160-4e06-a71c-d064acdffb19" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.387526] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 595.387702] env[62368]: DEBUG nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.387867] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 595.397472] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Releasing lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.397852] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.398055] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 595.398335] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-011956a2-dd0c-457c-89d8-1dc77cc3cb82 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.406574] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.410333] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076b55f2-5157-4069-841c-436343d97015 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.432665] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc could not be found. [ 595.432882] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 595.433072] env[62368]: INFO nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.433311] env[62368]: DEBUG oslo.service.loopingcall [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.435648] env[62368]: DEBUG nova.compute.manager [-] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.435749] env[62368]: DEBUG nova.network.neutron [-] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 595.449851] env[62368]: DEBUG nova.network.neutron [-] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.612756] env[62368]: DEBUG nova.compute.manager [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Received event network-changed-efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.612756] env[62368]: DEBUG nova.compute.manager [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Refreshing instance network info cache due to event network-changed-efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 595.612756] env[62368]: DEBUG oslo_concurrency.lockutils [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] Acquiring lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.612756] env[62368]: DEBUG oslo_concurrency.lockutils [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] Acquired lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.612756] env[62368]: DEBUG nova.network.neutron [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Refreshing network info cache for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 595.666192] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09554313-fb8f-4f7c-ba61-a8a3dfcc0f36 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.673835] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9e909c-e427-444d-ac3b-c73c10a9b9fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.703776] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d8a6f2-3e39-477a-9a42-9b5029d31c58 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.711016] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06ebe14-267b-4350-bfb3-1b8c28615754 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.723949] env[62368]: DEBUG nova.compute.provider_tree [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.922217] env[62368]: DEBUG nova.network.neutron [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.952020] env[62368]: DEBUG nova.network.neutron [-] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.130555] env[62368]: DEBUG nova.network.neutron [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.203529] env[62368]: DEBUG nova.network.neutron [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.226685] env[62368]: DEBUG nova.scheduler.client.report [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.425793] env[62368]: INFO nova.compute.manager [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] [instance: eac92d1f-d160-4e06-a71c-d064acdffb19] Took 1.04 seconds to deallocate network for instance. [ 596.454519] env[62368]: INFO nova.compute.manager [-] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Took 1.02 seconds to deallocate network for instance. [ 596.457034] env[62368]: DEBUG nova.compute.claims [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 596.457034] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.706362] env[62368]: DEBUG oslo_concurrency.lockutils [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] Releasing lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.706621] env[62368]: DEBUG nova.compute.manager [req-ad739612-41fe-4d12-96b3-b5cbdd7a1768 req-65fe406a-83a4-41ff-9127-3e3411945c0b service nova] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Received event network-vif-deleted-efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.732145] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.732631] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 596.735180] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.557s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.239908] env[62368]: DEBUG nova.compute.utils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 597.244824] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 597.244963] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 597.284818] env[62368]: DEBUG nova.policy [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53de6e2ad4d940ffa8063879d6c2d497', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '704ed0e906b447dd8ab4e3eea9ff14f7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 597.461404] env[62368]: INFO nova.scheduler.client.report [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Deleted allocations for instance eac92d1f-d160-4e06-a71c-d064acdffb19 [ 597.635516] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Successfully created port: 161c8b37-e02b-416e-ac52-5976f0eaaff5 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.646559] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe815dc6-e02f-4bcb-8eeb-241729ce1bb1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.655046] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791dd7c4-96ee-43eb-a467-f09ca897aca2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.686775] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22db8d21-7619-4a47-8356-dd813e59377e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.694444] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd01de0-e006-4d83-b74c-eb2b9b2172f8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.709057] env[62368]: DEBUG nova.compute.provider_tree [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.746057] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 597.971394] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e43fd56-1e6c-4424-b352-0007eceb0e01 tempest-VolumesAssistedSnapshotsTest-991320140 tempest-VolumesAssistedSnapshotsTest-991320140-project-member] Lock "eac92d1f-d160-4e06-a71c-d064acdffb19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.334s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.212966] env[62368]: DEBUG nova.scheduler.client.report [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.474928] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 598.635174] env[62368]: DEBUG nova.compute.manager [req-83735d9f-375a-425b-ba32-38b441750392 req-859697d4-e677-4894-93e3-5ded16d2fd2b service nova] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Received event network-changed-161c8b37-e02b-416e-ac52-5976f0eaaff5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.637264] env[62368]: DEBUG nova.compute.manager [req-83735d9f-375a-425b-ba32-38b441750392 req-859697d4-e677-4894-93e3-5ded16d2fd2b service nova] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Refreshing instance network info cache due to event network-changed-161c8b37-e02b-416e-ac52-5976f0eaaff5. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 598.637519] env[62368]: DEBUG oslo_concurrency.lockutils [req-83735d9f-375a-425b-ba32-38b441750392 req-859697d4-e677-4894-93e3-5ded16d2fd2b service nova] Acquiring lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.637664] env[62368]: DEBUG oslo_concurrency.lockutils [req-83735d9f-375a-425b-ba32-38b441750392 req-859697d4-e677-4894-93e3-5ded16d2fd2b service nova] Acquired lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.637825] env[62368]: DEBUG nova.network.neutron [req-83735d9f-375a-425b-ba32-38b441750392 req-859697d4-e677-4894-93e3-5ded16d2fd2b service nova] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Refreshing network info cache for port 161c8b37-e02b-416e-ac52-5976f0eaaff5 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 598.702306] env[62368]: ERROR nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. [ 598.702306] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.702306] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.702306] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.702306] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.702306] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.702306] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.702306] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.702306] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.702306] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 598.702306] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.702306] env[62368]: ERROR nova.compute.manager raise self.value [ 598.702306] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.702306] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.702306] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.702306] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.702889] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.702889] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.702889] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. [ 598.702889] env[62368]: ERROR nova.compute.manager [ 598.702889] env[62368]: Traceback (most recent call last): [ 598.702889] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.702889] env[62368]: listener.cb(fileno) [ 598.702889] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.702889] env[62368]: result = function(*args, **kwargs) [ 598.702889] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.702889] env[62368]: return func(*args, **kwargs) [ 598.702889] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.702889] env[62368]: raise e [ 598.702889] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.702889] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 598.702889] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.702889] env[62368]: created_port_ids = self._update_ports_for_instance( [ 598.702889] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.702889] env[62368]: with excutils.save_and_reraise_exception(): [ 598.702889] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.702889] env[62368]: self.force_reraise() [ 598.702889] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.702889] env[62368]: raise self.value [ 598.702889] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.702889] env[62368]: updated_port = self._update_port( [ 598.702889] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.702889] env[62368]: _ensure_no_port_binding_failure(port) [ 598.702889] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.702889] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.704627] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. [ 598.704627] env[62368]: Removing descriptor: 15 [ 598.721022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.721022] env[62368]: ERROR nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. [ 598.721022] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Traceback (most recent call last): [ 598.721022] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.721022] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self.driver.spawn(context, instance, image_meta, [ 598.721022] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.721022] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.721022] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.721022] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] vm_ref = self.build_virtual_machine(instance, [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] for vif in network_info: [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] return self._sync_wrapper(fn, *args, **kwargs) [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self.wait() [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self[:] = self._gt.wait() [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] return self._exit_event.wait() [ 598.721460] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] result = hub.switch() [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] return self.greenlet.switch() [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] result = function(*args, **kwargs) [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] return func(*args, **kwargs) [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] raise e [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] nwinfo = self.network_api.allocate_for_instance( [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.721913] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] created_port_ids = self._update_ports_for_instance( [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] with excutils.save_and_reraise_exception(): [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] self.force_reraise() [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] raise self.value [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] updated_port = self._update_port( [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] _ensure_no_port_binding_failure(port) [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.722386] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] raise exception.PortBindingFailed(port_id=port['id']) [ 598.722769] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] nova.exception.PortBindingFailed: Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. [ 598.722769] env[62368]: ERROR nova.compute.manager [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] [ 598.722769] env[62368]: DEBUG nova.compute.utils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 598.723362] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.293s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.727445] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Build of instance 01654798-4e54-4177-b20c-f6e0aecf72a7 was re-scheduled: Binding failed for port eade0b39-4a3a-4903-9ded-828f9f9ccc95, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 598.728557] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 598.728557] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Acquiring lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.728730] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Acquired lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.728950] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.756691] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 598.785220] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.785677] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.787561] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.787561] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.787561] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.787561] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.787561] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.787830] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.787830] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.787830] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.787830] env[62368]: DEBUG nova.virt.hardware [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.789335] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84ece2b-c47b-413d-a687-6a73078a0f56 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.798980] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a68f0c-46bf-4402-82f0-5a55316d81ac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.822145] env[62368]: ERROR nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Traceback (most recent call last): [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] yield resources [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self.driver.spawn(context, instance, image_meta, [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] vm_ref = self.build_virtual_machine(instance, [ 598.822145] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] for vif in network_info: [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] return self._sync_wrapper(fn, *args, **kwargs) [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self.wait() [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self[:] = self._gt.wait() [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] return self._exit_event.wait() [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 598.822582] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] current.throw(*self._exc) [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] result = function(*args, **kwargs) [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] return func(*args, **kwargs) [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] raise e [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] nwinfo = self.network_api.allocate_for_instance( [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] created_port_ids = self._update_ports_for_instance( [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] with excutils.save_and_reraise_exception(): [ 598.823212] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self.force_reraise() [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] raise self.value [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] updated_port = self._update_port( [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] _ensure_no_port_binding_failure(port) [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] raise exception.PortBindingFailed(port_id=port['id']) [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] nova.exception.PortBindingFailed: Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. [ 598.823776] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] [ 598.823776] env[62368]: INFO nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Terminating instance [ 598.824932] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Acquiring lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.998588] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.166063] env[62368]: DEBUG nova.network.neutron [req-83735d9f-375a-425b-ba32-38b441750392 req-859697d4-e677-4894-93e3-5ded16d2fd2b service nova] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.250518] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.275394] env[62368]: DEBUG nova.network.neutron [req-83735d9f-375a-425b-ba32-38b441750392 req-859697d4-e677-4894-93e3-5ded16d2fd2b service nova] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.336714] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.606136] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef45c9ef-3a6b-4133-b570-47a0ba4d743c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.614617] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cbf3ce-d4fe-454e-8775-5a94dba1434b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.645803] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bb1a21-6549-4b20-909f-924e669d142c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.653234] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93921c1c-bcd6-47aa-82e4-b97534433d26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.666486] env[62368]: DEBUG nova.compute.provider_tree [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.778454] env[62368]: DEBUG oslo_concurrency.lockutils [req-83735d9f-375a-425b-ba32-38b441750392 req-859697d4-e677-4894-93e3-5ded16d2fd2b service nova] Releasing lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.778744] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Acquired lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.778930] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.839394] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Releasing lock "refresh_cache-01654798-4e54-4177-b20c-f6e0aecf72a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.839651] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 599.839846] env[62368]: DEBUG nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.840030] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.859603] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.169087] env[62368]: DEBUG nova.scheduler.client.report [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.300038] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.362745] env[62368]: DEBUG nova.network.neutron [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.605976] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.656468] env[62368]: DEBUG nova.compute.manager [req-a1cf85e1-29df-424a-b919-e9ef26f976d4 req-31f80949-38a8-4d5a-93ab-a2277c7a45d3 service nova] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Received event network-vif-deleted-161c8b37-e02b-416e-ac52-5976f0eaaff5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.674580] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.675202] env[62368]: ERROR nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Traceback (most recent call last): [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self.driver.spawn(context, instance, image_meta, [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] vm_ref = self.build_virtual_machine(instance, [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.675202] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] for vif in network_info: [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] return self._sync_wrapper(fn, *args, **kwargs) [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self.wait() [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self[:] = self._gt.wait() [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] return self._exit_event.wait() [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] result = hub.switch() [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.675680] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] return self.greenlet.switch() [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] result = function(*args, **kwargs) [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] return func(*args, **kwargs) [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] raise e [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] nwinfo = self.network_api.allocate_for_instance( [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] created_port_ids = self._update_ports_for_instance( [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] with excutils.save_and_reraise_exception(): [ 600.676349] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] self.force_reraise() [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] raise self.value [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] updated_port = self._update_port( [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] _ensure_no_port_binding_failure(port) [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] raise exception.PortBindingFailed(port_id=port['id']) [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] nova.exception.PortBindingFailed: Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. [ 600.676890] env[62368]: ERROR nova.compute.manager [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] [ 600.677427] env[62368]: DEBUG nova.compute.utils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 600.677427] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.186s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.678419] env[62368]: INFO nova.compute.claims [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.681212] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Build of instance c9fe2763-2d3c-48f6-a836-e0c8634acc95 was re-scheduled: Binding failed for port ef23e468-378f-4027-8c2c-94380011a2a8, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 600.681687] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 600.681893] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.682053] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquired lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.682212] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.864994] env[62368]: INFO nova.compute.manager [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] [instance: 01654798-4e54-4177-b20c-f6e0aecf72a7] Took 1.02 seconds to deallocate network for instance. [ 601.109254] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Releasing lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.109689] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.109879] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 601.110210] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10fe2430-eb78-422e-a390-917b86bd440f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.119354] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b16c178-9650-47f4-ab73-28f765b4cdfb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.143028] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1523daf8-033b-48e2-a495-885feb4e0ea3 could not be found. [ 601.145345] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 601.145345] env[62368]: INFO nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 601.145345] env[62368]: DEBUG oslo.service.loopingcall [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.145345] env[62368]: DEBUG nova.compute.manager [-] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.145345] env[62368]: DEBUG nova.network.neutron [-] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 601.161280] env[62368]: DEBUG nova.network.neutron [-] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.211451] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.257761] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.666034] env[62368]: DEBUG nova.network.neutron [-] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.761934] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Releasing lock "refresh_cache-c9fe2763-2d3c-48f6-a836-e0c8634acc95" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.761934] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 601.761934] env[62368]: DEBUG nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.761934] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 601.791123] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.896340] env[62368]: INFO nova.scheduler.client.report [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Deleted allocations for instance 01654798-4e54-4177-b20c-f6e0aecf72a7 [ 602.168596] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79307c8-1579-4efe-a1fc-e39085d30a74 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.169599] env[62368]: INFO nova.compute.manager [-] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Took 1.03 seconds to deallocate network for instance. [ 602.171859] env[62368]: DEBUG nova.compute.claims [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.174905] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.177747] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7211984d-55af-475b-8119-3e30b6e55b32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.207642] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c411d86b-3b58-4522-8a7f-dd4aa0c99a45 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.215749] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c06d6e-3c2d-416f-a7dc-d12c350b669c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.232021] env[62368]: DEBUG nova.compute.provider_tree [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.293148] env[62368]: DEBUG nova.network.neutron [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.407168] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9851c6e8-0cfe-4bf4-8d93-6087d10f8ed6 tempest-ServersV294TestFqdnHostnames-2029629489 tempest-ServersV294TestFqdnHostnames-2029629489-project-member] Lock "01654798-4e54-4177-b20c-f6e0aecf72a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.924s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.735025] env[62368]: DEBUG nova.scheduler.client.report [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.795707] env[62368]: INFO nova.compute.manager [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: c9fe2763-2d3c-48f6-a836-e0c8634acc95] Took 1.03 seconds to deallocate network for instance. [ 602.908446] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.239413] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.240041] env[62368]: DEBUG nova.compute.manager [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.243230] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.219s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.440743] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.747806] env[62368]: DEBUG nova.compute.utils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.749279] env[62368]: DEBUG nova.compute.manager [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Not allocating networking since 'none' was specified. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 603.839353] env[62368]: INFO nova.scheduler.client.report [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Deleted allocations for instance c9fe2763-2d3c-48f6-a836-e0c8634acc95 [ 604.166424] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eee8c6-1808-48c0-a2ac-c4a1855aee3e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.174158] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7d1013-713b-4ee9-a86d-c181647695aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.206839] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b559206-6c76-4a10-9cd3-5043edc75afc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.214407] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f270b9-4ffb-4367-99c3-8df388f0c3e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.227312] env[62368]: DEBUG nova.compute.provider_tree [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.255121] env[62368]: DEBUG nova.compute.manager [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.350156] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2685a20c-0a68-47b5-b78f-6fa260e0bae1 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "c9fe2763-2d3c-48f6-a836-e0c8634acc95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.642s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.731411] env[62368]: DEBUG nova.scheduler.client.report [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.852595] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.237256] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.237443] env[62368]: ERROR nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Traceback (most recent call last): [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self.driver.spawn(context, instance, image_meta, [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] vm_ref = self.build_virtual_machine(instance, [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.237443] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] for vif in network_info: [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] return self._sync_wrapper(fn, *args, **kwargs) [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self.wait() [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self[:] = self._gt.wait() [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] return self._exit_event.wait() [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] result = hub.switch() [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.237954] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] return self.greenlet.switch() [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] result = function(*args, **kwargs) [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] return func(*args, **kwargs) [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] raise e [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] nwinfo = self.network_api.allocate_for_instance( [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] created_port_ids = self._update_ports_for_instance( [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] with excutils.save_and_reraise_exception(): [ 605.238358] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] self.force_reraise() [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] raise self.value [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] updated_port = self._update_port( [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] _ensure_no_port_binding_failure(port) [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] raise exception.PortBindingFailed(port_id=port['id']) [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] nova.exception.PortBindingFailed: Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. [ 605.238718] env[62368]: ERROR nova.compute.manager [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] [ 605.239012] env[62368]: DEBUG nova.compute.utils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.239862] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.624s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.246775] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Build of instance 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23 was re-scheduled: Binding failed for port ee270188-766a-470c-bc9d-ac8f401df526, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 605.246775] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 605.246920] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Acquiring lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.247047] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Acquired lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.247245] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.265312] env[62368]: DEBUG nova.compute.manager [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.305535] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.305535] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.305535] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.305712] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.305712] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.305712] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.305712] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.305712] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.305925] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.305925] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.305925] env[62368]: DEBUG nova.virt.hardware [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.306714] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7166e8f3-fda5-40b0-ae6f-2993717ceaa3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.317279] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53053dc-9c34-4405-9848-67bfdebae5d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.333817] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 605.340211] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Creating folder: Project (d1ac177b76d549cda71d0c76556aee72). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 605.340553] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f21298ee-825a-4a3b-a37c-211e5434d026 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.351050] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Created folder: Project (d1ac177b76d549cda71d0c76556aee72) in parent group-v259706. [ 605.351255] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Creating folder: Instances. Parent ref: group-v259716. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 605.351598] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3be391a-9439-4d05-ac78-5ff661355dbc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.361422] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Created folder: Instances in parent group-v259716. [ 605.361741] env[62368]: DEBUG oslo.service.loopingcall [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.363996] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 605.364435] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0f18325-28c0-45d7-9fce-c33d46bf885a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.381739] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 605.381739] env[62368]: value = "task-1198162" [ 605.381739] env[62368]: _type = "Task" [ 605.381739] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.386938] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.390052] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198162, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.782604] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.893210] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.900320] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198162, 'name': CreateVM_Task, 'duration_secs': 0.265471} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.903041] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 605.903680] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.904152] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.904323] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 605.905731] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6048923b-6477-4137-a5f9-3d801f84183b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.910599] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 605.910599] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52223011-5936-b30b-de23-3d36e7c43766" [ 605.910599] env[62368]: _type = "Task" [ 605.910599] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.923673] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52223011-5936-b30b-de23-3d36e7c43766, 'name': SearchDatastore_Task, 'duration_secs': 0.009211} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.923673] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.923673] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 605.923958] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.923958] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.924040] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 605.924625] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d47b0a4f-e45f-43a7-bcfc-b6624628ab52 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.935026] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 605.935026] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 605.935026] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4af38718-3611-4b9d-9b6e-467ec923cda8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.938972] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 605.938972] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52209c79-a91c-af61-56ea-04d55d40af53" [ 605.938972] env[62368]: _type = "Task" [ 605.938972] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.951756] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52209c79-a91c-af61-56ea-04d55d40af53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.177453] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c39a7ac-cca0-48c4-bc89-f34a289bed7f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.180515] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.180515] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.186820] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a30c3dc-29bc-4923-93b3-981882af075c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.219877] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927d78e6-6cab-484c-b28d-061d083791f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.227017] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d532d2b9-d1a8-45ee-8208-cc3060d03f4f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.242198] env[62368]: DEBUG nova.compute.provider_tree [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.404377] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Releasing lock "refresh_cache-15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.404610] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.404793] env[62368]: DEBUG nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.404958] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.422900] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.453964] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52209c79-a91c-af61-56ea-04d55d40af53, 'name': SearchDatastore_Task, 'duration_secs': 0.008546} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.455018] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee00fc02-7b8b-4b00-a70d-1bf4440e4c2a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.461977] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 606.461977] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522d4058-a999-2d35-ea7f-853391ffb45c" [ 606.461977] env[62368]: _type = "Task" [ 606.461977] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.469395] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522d4058-a999-2d35-ea7f-853391ffb45c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.691851] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.691851] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 606.691970] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Rebuilding the list of instances to heal {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 606.746979] env[62368]: DEBUG nova.scheduler.client.report [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.925591] env[62368]: DEBUG nova.network.neutron [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.974120] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522d4058-a999-2d35-ea7f-853391ffb45c, 'name': SearchDatastore_Task, 'duration_secs': 0.008534} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.974437] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.974699] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] d22aff27-739f-4b0d-a0e4-b6316c252c3a/d22aff27-739f-4b0d-a0e4-b6316c252c3a.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 606.974950] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3586bfee-db72-48ad-a2e4-e980a26f21b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.981304] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 606.981304] env[62368]: value = "task-1198163" [ 606.981304] env[62368]: _type = "Task" [ 606.981304] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.993029] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198163, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.196666] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 607.196989] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 607.196989] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 607.197096] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 607.197616] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 607.197616] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Didn't find any instances for network info cache update. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 607.197616] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.197736] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.197913] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.198091] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.198239] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.198909] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.198909] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 607.198909] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.252636] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.253321] env[62368]: ERROR nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Traceback (most recent call last): [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self.driver.spawn(context, instance, image_meta, [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] vm_ref = self.build_virtual_machine(instance, [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.253321] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] for vif in network_info: [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] return self._sync_wrapper(fn, *args, **kwargs) [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self.wait() [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self[:] = self._gt.wait() [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] return self._exit_event.wait() [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] current.throw(*self._exc) [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.253853] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] result = function(*args, **kwargs) [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] return func(*args, **kwargs) [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] raise e [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] nwinfo = self.network_api.allocate_for_instance( [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] created_port_ids = self._update_ports_for_instance( [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] with excutils.save_and_reraise_exception(): [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] self.force_reraise() [ 607.254376] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] raise self.value [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] updated_port = self._update_port( [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] _ensure_no_port_binding_failure(port) [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] raise exception.PortBindingFailed(port_id=port['id']) [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] nova.exception.PortBindingFailed: Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. [ 607.254722] env[62368]: ERROR nova.compute.manager [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] [ 607.254722] env[62368]: DEBUG nova.compute.utils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.255778] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.452s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.258719] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Build of instance 546a40aa-c278-460c-8c91-c6c903d751ab was re-scheduled: Binding failed for port 80af8aeb-f5f2-48a3-a218-8f0fab2efc6a, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 607.259193] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 607.259424] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Acquiring lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.259569] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Acquired lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.259847] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.428590] env[62368]: INFO nova.compute.manager [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] [instance: 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23] Took 1.02 seconds to deallocate network for instance. [ 607.493235] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198163, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482701} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.493659] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] d22aff27-739f-4b0d-a0e4-b6316c252c3a/d22aff27-739f-4b0d-a0e4-b6316c252c3a.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 607.493919] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 607.494221] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ef895b7-768a-45c9-a6e8-370880c277d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.501682] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 607.501682] env[62368]: value = "task-1198164" [ 607.501682] env[62368]: _type = "Task" [ 607.501682] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.510673] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198164, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.704143] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.791774] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.957548] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.011948] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198164, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062209} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.016556] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 608.021225] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e74a3d9-5dc2-4cdf-9432-b5a42b067f23 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.046940] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Reconfiguring VM instance instance-00000015 to attach disk [datastore1] d22aff27-739f-4b0d-a0e4-b6316c252c3a/d22aff27-739f-4b0d-a0e4-b6316c252c3a.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 608.050582] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df9336e0-a73b-4d27-b691-df186d375d13 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.072764] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 608.072764] env[62368]: value = "task-1198165" [ 608.072764] env[62368]: _type = "Task" [ 608.072764] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.081254] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.304606] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b85008f-8db8-410d-ae89-e08c4585e63f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.311277] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bae9168-e7c2-4915-9846-fcc6446f51d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.342724] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac109238-4cf0-4a12-9a20-ea0666bfc0f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.350439] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cd27c3-c07a-49d5-83a7-69e711352083 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.365852] env[62368]: DEBUG nova.compute.provider_tree [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.467492] env[62368]: INFO nova.scheduler.client.report [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Deleted allocations for instance 15f43ac0-ef8c-480c-bc07-05d2dbf0ee23 [ 608.475341] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Releasing lock "refresh_cache-546a40aa-c278-460c-8c91-c6c903d751ab" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.475753] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 608.475977] env[62368]: DEBUG nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.476158] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.500843] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.586247] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198165, 'name': ReconfigVM_Task, 'duration_secs': 0.286766} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.586516] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Reconfigured VM instance instance-00000015 to attach disk [datastore1] d22aff27-739f-4b0d-a0e4-b6316c252c3a/d22aff27-739f-4b0d-a0e4-b6316c252c3a.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 608.587215] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c248f12-4f59-4d67-9318-8166adcd7374 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.593511] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 608.593511] env[62368]: value = "task-1198166" [ 608.593511] env[62368]: _type = "Task" [ 608.593511] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.602544] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198166, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.870403] env[62368]: DEBUG nova.scheduler.client.report [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.982305] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b8aa88d9-2e8c-4bc4-85c2-87154bad12bf tempest-ServersWithSpecificFlavorTestJSON-1975185740 tempest-ServersWithSpecificFlavorTestJSON-1975185740-project-member] Lock "15f43ac0-ef8c-480c-bc07-05d2dbf0ee23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.394s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.003393] env[62368]: DEBUG nova.network.neutron [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.104488] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198166, 'name': Rename_Task, 'duration_secs': 0.153335} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.104804] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 609.105073] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9ded0c4-c660-4d92-9b9e-8245486db8df {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.112682] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 609.112682] env[62368]: value = "task-1198167" [ 609.112682] env[62368]: _type = "Task" [ 609.112682] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.121115] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198167, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.375456] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.120s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.376068] env[62368]: ERROR nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Traceback (most recent call last): [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self.driver.spawn(context, instance, image_meta, [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] vm_ref = self.build_virtual_machine(instance, [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.376068] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] for vif in network_info: [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] return self._sync_wrapper(fn, *args, **kwargs) [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self.wait() [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self[:] = self._gt.wait() [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] return self._exit_event.wait() [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] result = hub.switch() [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.376492] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] return self.greenlet.switch() [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] result = function(*args, **kwargs) [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] return func(*args, **kwargs) [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] raise e [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] nwinfo = self.network_api.allocate_for_instance( [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] created_port_ids = self._update_ports_for_instance( [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] with excutils.save_and_reraise_exception(): [ 609.376792] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] self.force_reraise() [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] raise self.value [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] updated_port = self._update_port( [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] _ensure_no_port_binding_failure(port) [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] raise exception.PortBindingFailed(port_id=port['id']) [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] nova.exception.PortBindingFailed: Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. [ 609.377111] env[62368]: ERROR nova.compute.manager [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] [ 609.377377] env[62368]: DEBUG nova.compute.utils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.379802] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Build of instance fd5478d6-9fa8-4471-8b95-b89f3d826bcc was re-scheduled: Binding failed for port bc51ee6a-376f-4ffc-be51-e6e61d8acca9, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.380237] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.380505] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquiring lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.380661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Acquired lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.380849] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.384680] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.481s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.468702] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquiring lock "3f36000e-b93d-4dda-ac39-b8459203c227" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.468935] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "3f36000e-b93d-4dda-ac39-b8459203c227" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.484683] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.507750] env[62368]: INFO nova.compute.manager [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] [instance: 546a40aa-c278-460c-8c91-c6c903d751ab] Took 1.03 seconds to deallocate network for instance. [ 609.624135] env[62368]: DEBUG oslo_vmware.api [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198167, 'name': PowerOnVM_Task, 'duration_secs': 0.454436} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.624135] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 609.624219] env[62368]: INFO nova.compute.manager [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Took 4.36 seconds to spawn the instance on the hypervisor. [ 609.626169] env[62368]: DEBUG nova.compute.manager [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 609.626169] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cabbcea-6c79-4ebc-8483-1ae19381e688 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.927212] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.022807] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.055301] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.146520] env[62368]: INFO nova.compute.manager [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Took 27.67 seconds to build instance. [ 610.803786] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Releasing lock "refresh_cache-fd5478d6-9fa8-4471-8b95-b89f3d826bcc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.804012] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.804309] env[62368]: DEBUG nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.804384] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.805887] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbb9724b-96c6-451c-b89b-7e580f1c7dde tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.221s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.807742] env[62368]: DEBUG nova.compute.manager [None req-b5134c3e-f840-4dd6-98c0-1bcfa803b60e tempest-ServerDiagnosticsV248Test-1074996794 tempest-ServerDiagnosticsV248Test-1074996794-project-admin] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 610.811858] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8444e4d7-6127-4d9b-8b0f-24f6501e87e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.817152] env[62368]: INFO nova.compute.manager [None req-b5134c3e-f840-4dd6-98c0-1bcfa803b60e tempest-ServerDiagnosticsV248Test-1074996794 tempest-ServerDiagnosticsV248Test-1074996794-project-admin] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Retrieving diagnostics [ 610.817995] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05bb9c47-4f20-4b97-8de8-0224218f5402 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.850144] env[62368]: INFO nova.scheduler.client.report [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Deleted allocations for instance 546a40aa-c278-460c-8c91-c6c903d751ab [ 610.855993] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.021015] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbfcba2-eac9-4bd7-a205-47b2480efc4a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.028950] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d23cfe7-7b8a-4945-bba0-5e11771aad2c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.061578] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513f0891-1b3a-4580-be08-1842dc966822 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.069466] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ff32ca-9a63-4aac-9c54-bfb93f2de925 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.086844] env[62368]: DEBUG nova.compute.provider_tree [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.313402] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.360522] env[62368]: DEBUG nova.network.neutron [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.363154] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79061a50-0204-49f2-92ae-c34a2f3d32dd tempest-ServersTestBootFromVolume-869106810 tempest-ServersTestBootFromVolume-869106810-project-member] Lock "546a40aa-c278-460c-8c91-c6c903d751ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.879s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.593835] env[62368]: DEBUG nova.scheduler.client.report [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.847240] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.864272] env[62368]: INFO nova.compute.manager [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] [instance: fd5478d6-9fa8-4471-8b95-b89f3d826bcc] Took 1.06 seconds to deallocate network for instance. [ 611.867269] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 612.100198] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.718s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.100815] env[62368]: ERROR nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Traceback (most recent call last): [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self.driver.spawn(context, instance, image_meta, [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] vm_ref = self.build_virtual_machine(instance, [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.100815] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] for vif in network_info: [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] return self._sync_wrapper(fn, *args, **kwargs) [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self.wait() [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self[:] = self._gt.wait() [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] return self._exit_event.wait() [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] current.throw(*self._exc) [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.101114] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] result = function(*args, **kwargs) [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] return func(*args, **kwargs) [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] raise e [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] nwinfo = self.network_api.allocate_for_instance( [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] created_port_ids = self._update_ports_for_instance( [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] with excutils.save_and_reraise_exception(): [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] self.force_reraise() [ 612.101414] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] raise self.value [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] updated_port = self._update_port( [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] _ensure_no_port_binding_failure(port) [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] raise exception.PortBindingFailed(port_id=port['id']) [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] nova.exception.PortBindingFailed: Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. [ 612.101685] env[62368]: ERROR nova.compute.manager [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] [ 612.101685] env[62368]: DEBUG nova.compute.utils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.103197] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.646s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.108029] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Build of instance fc4267ff-dda1-444f-a5b0-a4954626b42d was re-scheduled: Binding failed for port 38d0cf3e-d303-4caf-8bc9-0895237d59a1, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.108029] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.108029] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.108029] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquired lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.108237] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 612.395231] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.631966] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.632235] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.635752] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.779753] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.913520] env[62368]: INFO nova.scheduler.client.report [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Deleted allocations for instance fd5478d6-9fa8-4471-8b95-b89f3d826bcc [ 613.088174] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6936ff-0086-427e-9e41-7793dffa12d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.097674] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7eb5e99-c24b-43f1-8b3c-ca1647f5c3b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.136235] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542327dc-e049-43f6-b18a-d20975b922d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.145485] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1612903-6cc4-4689-8261-f5897452faee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.160106] env[62368]: DEBUG nova.compute.provider_tree [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.285087] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Releasing lock "refresh_cache-fc4267ff-dda1-444f-a5b0-a4954626b42d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.285087] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.285087] env[62368]: DEBUG nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.285087] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 613.323885] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.423071] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c088abb2-d19d-4df4-8c25-3d05a6318676 tempest-ListImageFiltersTestJSON-2028716966 tempest-ListImageFiltersTestJSON-2028716966-project-member] Lock "fd5478d6-9fa8-4471-8b95-b89f3d826bcc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.411s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.663985] env[62368]: DEBUG nova.scheduler.client.report [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.827379] env[62368]: DEBUG nova.network.neutron [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.927302] env[62368]: DEBUG nova.compute.manager [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.171460] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.068s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.172626] env[62368]: ERROR nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Traceback (most recent call last): [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self.driver.spawn(context, instance, image_meta, [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] vm_ref = self.build_virtual_machine(instance, [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.172626] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] for vif in network_info: [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] return self._sync_wrapper(fn, *args, **kwargs) [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self.wait() [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self[:] = self._gt.wait() [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] return self._exit_event.wait() [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] current.throw(*self._exc) [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.173628] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] result = function(*args, **kwargs) [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] return func(*args, **kwargs) [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] raise e [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] nwinfo = self.network_api.allocate_for_instance( [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] created_port_ids = self._update_ports_for_instance( [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] with excutils.save_and_reraise_exception(): [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] self.force_reraise() [ 614.174248] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] raise self.value [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] updated_port = self._update_port( [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] _ensure_no_port_binding_failure(port) [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] raise exception.PortBindingFailed(port_id=port['id']) [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] nova.exception.PortBindingFailed: Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. [ 614.174777] env[62368]: ERROR nova.compute.manager [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] [ 614.174777] env[62368]: DEBUG nova.compute.utils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.175080] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.176s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.180271] env[62368]: INFO nova.compute.claims [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.183009] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Build of instance 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc was re-scheduled: Binding failed for port efdaf1ac-f10a-49c1-95a2-07e1cbd8d1bd, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 614.183456] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 614.183764] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Acquiring lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.183847] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Acquired lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.183958] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.331312] env[62368]: INFO nova.compute.manager [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: fc4267ff-dda1-444f-a5b0-a4954626b42d] Took 1.05 seconds to deallocate network for instance. [ 614.457556] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.706663] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.785586] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.288429] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Releasing lock "refresh_cache-489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.290214] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.290214] env[62368]: DEBUG nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.290214] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.305666] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.368275] env[62368]: INFO nova.scheduler.client.report [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Deleted allocations for instance fc4267ff-dda1-444f-a5b0-a4954626b42d [ 615.605623] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c86f20e-0434-4139-9685-402714dabf83 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.613379] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b3cd16-ee70-4b1a-85f8-b307dd6888ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.643972] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efbdc25-9dda-4d2d-9d95-f574576d7945 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.651243] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07854189-11c9-449b-8e63-19f58d531fa6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.664427] env[62368]: DEBUG nova.compute.provider_tree [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.811067] env[62368]: DEBUG nova.network.neutron [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.882042] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0c398200-21a9-4c8d-9e15-5ec433212a94 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "fc4267ff-dda1-444f-a5b0-a4954626b42d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.694s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.169884] env[62368]: DEBUG nova.scheduler.client.report [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.315136] env[62368]: INFO nova.compute.manager [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] [instance: 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc] Took 1.03 seconds to deallocate network for instance. [ 616.386367] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.676871] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.677840] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.680732] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.508s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.912796] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.188010] env[62368]: DEBUG nova.compute.utils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.195719] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.195935] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 617.287128] env[62368]: DEBUG nova.policy [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f3afe492e7e47f28c3c8dad61f819d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81b69987192149ddb6dfcafc1f7770fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.356752] env[62368]: INFO nova.scheduler.client.report [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Deleted allocations for instance 489cd441-f925-4cfd-8a6a-9f4d3e22d6dc [ 617.669230] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa749fd-b826-4508-b0b7-559e25e31b44 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.677022] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e432cf-5e52-4418-97fd-ab9db1bc9811 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.712871] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.718024] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cc1574-bf6f-4661-a79a-139791240146 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.725639] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb55fbd-76a9-4802-bea6-092a25562820 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.739819] env[62368]: DEBUG nova.compute.provider_tree [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.790715] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Successfully created port: 1c9033c1-2010-4274-82e7-7d886156b319 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.802400] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.802400] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.872978] env[62368]: DEBUG oslo_concurrency.lockutils [None req-319f60a0-c9e6-4dc1-b178-79caff7ac6ec tempest-TenantUsagesTestJSON-751618129 tempest-TenantUsagesTestJSON-751618129-project-member] Lock "489cd441-f925-4cfd-8a6a-9f4d3e22d6dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.756s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.244978] env[62368]: DEBUG nova.scheduler.client.report [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.378980] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.733400] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.752983] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.753826] env[62368]: ERROR nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Traceback (most recent call last): [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self.driver.spawn(context, instance, image_meta, [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] vm_ref = self.build_virtual_machine(instance, [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.753826] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] for vif in network_info: [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] return self._sync_wrapper(fn, *args, **kwargs) [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self.wait() [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self[:] = self._gt.wait() [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] return self._exit_event.wait() [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] current.throw(*self._exc) [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.754127] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] result = function(*args, **kwargs) [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] return func(*args, **kwargs) [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] raise e [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] nwinfo = self.network_api.allocate_for_instance( [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] created_port_ids = self._update_ports_for_instance( [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] with excutils.save_and_reraise_exception(): [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] self.force_reraise() [ 618.754430] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] raise self.value [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] updated_port = self._update_port( [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] _ensure_no_port_binding_failure(port) [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] raise exception.PortBindingFailed(port_id=port['id']) [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] nova.exception.PortBindingFailed: Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. [ 618.754718] env[62368]: ERROR nova.compute.manager [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] [ 618.754718] env[62368]: DEBUG nova.compute.utils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.758545] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Build of instance 1523daf8-033b-48e2-a495-885feb4e0ea3 was re-scheduled: Binding failed for port 161c8b37-e02b-416e-ac52-5976f0eaaff5, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 618.758545] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 618.758545] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Acquiring lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.758835] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Acquired lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.758994] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.764030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.323s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.765802] env[62368]: INFO nova.compute.claims [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.779350] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.779903] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.779903] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.780015] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.781149] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.781343] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.781563] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.781995] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.781995] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.782106] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.782288] env[62368]: DEBUG nova.virt.hardware [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.783677] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b20e64-8e1b-489a-861e-d1c0b8f9f22f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.794860] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feaaca0d-9bea-41d0-a13f-d33304c2cd57 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.907414] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.304285] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.392585] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.523732] env[62368]: ERROR nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. [ 619.523732] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 619.523732] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.523732] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 619.523732] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.523732] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 619.523732] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.523732] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 619.523732] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.523732] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 619.523732] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.523732] env[62368]: ERROR nova.compute.manager raise self.value [ 619.523732] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.523732] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 619.523732] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.523732] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 619.524164] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.524164] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 619.524164] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. [ 619.524164] env[62368]: ERROR nova.compute.manager [ 619.524164] env[62368]: Traceback (most recent call last): [ 619.524164] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 619.524164] env[62368]: listener.cb(fileno) [ 619.524164] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.524164] env[62368]: result = function(*args, **kwargs) [ 619.524164] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.524164] env[62368]: return func(*args, **kwargs) [ 619.524164] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.524164] env[62368]: raise e [ 619.524164] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.524164] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 619.524164] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.524164] env[62368]: created_port_ids = self._update_ports_for_instance( [ 619.524164] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.524164] env[62368]: with excutils.save_and_reraise_exception(): [ 619.524164] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.524164] env[62368]: self.force_reraise() [ 619.524164] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.524164] env[62368]: raise self.value [ 619.524164] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.524164] env[62368]: updated_port = self._update_port( [ 619.524164] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.524164] env[62368]: _ensure_no_port_binding_failure(port) [ 619.524164] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.524164] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 619.524812] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. [ 619.524812] env[62368]: Removing descriptor: 18 [ 619.524812] env[62368]: ERROR nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Traceback (most recent call last): [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] yield resources [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self.driver.spawn(context, instance, image_meta, [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.524812] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] vm_ref = self.build_virtual_machine(instance, [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] for vif in network_info: [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] return self._sync_wrapper(fn, *args, **kwargs) [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self.wait() [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self[:] = self._gt.wait() [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] return self._exit_event.wait() [ 619.525102] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] result = hub.switch() [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] return self.greenlet.switch() [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] result = function(*args, **kwargs) [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] return func(*args, **kwargs) [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] raise e [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] nwinfo = self.network_api.allocate_for_instance( [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.525401] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] created_port_ids = self._update_ports_for_instance( [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] with excutils.save_and_reraise_exception(): [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self.force_reraise() [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] raise self.value [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] updated_port = self._update_port( [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] _ensure_no_port_binding_failure(port) [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.525794] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] raise exception.PortBindingFailed(port_id=port['id']) [ 619.526107] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] nova.exception.PortBindingFailed: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. [ 619.526107] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] [ 619.526107] env[62368]: INFO nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Terminating instance [ 619.531502] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.531502] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.531502] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 619.683898] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquiring lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.683898] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.686300] env[62368]: DEBUG nova.compute.manager [req-5c4f864b-3f33-4587-aacf-d6850dffc3b0 req-d34d2955-3afc-4d75-829e-c95e7417d869 service nova] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Received event network-changed-1c9033c1-2010-4274-82e7-7d886156b319 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.686482] env[62368]: DEBUG nova.compute.manager [req-5c4f864b-3f33-4587-aacf-d6850dffc3b0 req-d34d2955-3afc-4d75-829e-c95e7417d869 service nova] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Refreshing instance network info cache due to event network-changed-1c9033c1-2010-4274-82e7-7d886156b319. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 619.686656] env[62368]: DEBUG oslo_concurrency.lockutils [req-5c4f864b-3f33-4587-aacf-d6850dffc3b0 req-d34d2955-3afc-4d75-829e-c95e7417d869 service nova] Acquiring lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.895404] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Releasing lock "refresh_cache-1523daf8-033b-48e2-a495-885feb4e0ea3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.897029] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 619.897029] env[62368]: DEBUG nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.897029] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.928058] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.060304] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.215924] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.283223] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6945a569-c630-4cdf-aef1-f0e0a97946a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.291288] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b150be-f06f-4585-80e5-0a852105fe5a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.322640] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee2078b-ada7-4aaa-b821-b5624f0a7ce1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.330782] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487332ea-a1b4-4f3b-8660-f3d13f21867e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.344779] env[62368]: DEBUG nova.compute.provider_tree [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.356328] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquiring lock "b393f654-0611-493e-a89f-445dfd3f8c13" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.356432] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "b393f654-0611-493e-a89f-445dfd3f8c13" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.430642] env[62368]: DEBUG nova.network.neutron [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.719583] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.720031] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.720289] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 620.722492] env[62368]: DEBUG oslo_concurrency.lockutils [req-5c4f864b-3f33-4587-aacf-d6850dffc3b0 req-d34d2955-3afc-4d75-829e-c95e7417d869 service nova] Acquired lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.722492] env[62368]: DEBUG nova.network.neutron [req-5c4f864b-3f33-4587-aacf-d6850dffc3b0 req-d34d2955-3afc-4d75-829e-c95e7417d869 service nova] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Refreshing network info cache for port 1c9033c1-2010-4274-82e7-7d886156b319 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 620.722492] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f79f3594-313d-4823-a6bb-1d6d783f92bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.733701] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12d8c1a-b0b9-4883-9d90-3b25c81f4603 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.761292] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c179402-a881-4b89-89d7-5e9031ba78cf could not be found. [ 620.761576] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 620.761766] env[62368]: INFO nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 620.762435] env[62368]: DEBUG oslo.service.loopingcall [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.762707] env[62368]: DEBUG nova.compute.manager [-] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.762912] env[62368]: DEBUG nova.network.neutron [-] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 620.789646] env[62368]: DEBUG nova.network.neutron [-] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.850858] env[62368]: DEBUG nova.scheduler.client.report [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.934119] env[62368]: INFO nova.compute.manager [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] [instance: 1523daf8-033b-48e2-a495-885feb4e0ea3] Took 1.04 seconds to deallocate network for instance. [ 621.108364] env[62368]: DEBUG nova.compute.manager [None req-6b58f050-8acb-4e44-bfea-1031600d4ae4 tempest-ServerDiagnosticsV248Test-1074996794 tempest-ServerDiagnosticsV248Test-1074996794-project-admin] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 621.109511] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f92083-c7a4-456c-9afc-858c56c39457 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.117227] env[62368]: INFO nova.compute.manager [None req-6b58f050-8acb-4e44-bfea-1031600d4ae4 tempest-ServerDiagnosticsV248Test-1074996794 tempest-ServerDiagnosticsV248Test-1074996794-project-admin] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Retrieving diagnostics [ 621.118207] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a943610-3967-439f-8b11-3cef2cee5ca4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.246556] env[62368]: DEBUG nova.network.neutron [req-5c4f864b-3f33-4587-aacf-d6850dffc3b0 req-d34d2955-3afc-4d75-829e-c95e7417d869 service nova] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.292535] env[62368]: DEBUG nova.network.neutron [-] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.356479] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.357013] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.360451] env[62368]: DEBUG nova.network.neutron [req-5c4f864b-3f33-4587-aacf-d6850dffc3b0 req-d34d2955-3afc-4d75-829e-c95e7417d869 service nova] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.361451] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.975s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.362922] env[62368]: INFO nova.compute.claims [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.795199] env[62368]: INFO nova.compute.manager [-] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Took 1.03 seconds to deallocate network for instance. [ 621.797944] env[62368]: DEBUG nova.compute.claims [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.797944] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.851667] env[62368]: DEBUG nova.compute.manager [req-07c3f06a-533c-4fc9-b89c-358938b92b09 req-8828b0c0-66be-4bf9-a8c8-0af14d11119e service nova] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Received event network-vif-deleted-1c9033c1-2010-4274-82e7-7d886156b319 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.865795] env[62368]: DEBUG oslo_concurrency.lockutils [req-5c4f864b-3f33-4587-aacf-d6850dffc3b0 req-d34d2955-3afc-4d75-829e-c95e7417d869 service nova] Releasing lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.867120] env[62368]: DEBUG nova.compute.utils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.871027] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.871306] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.929808] env[62368]: DEBUG nova.policy [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ec6fb90433943ca9711d4a7a5e40bef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd4cd004b0d54bb49a88722fd1312249', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.994045] env[62368]: INFO nova.scheduler.client.report [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Deleted allocations for instance 1523daf8-033b-48e2-a495-885feb4e0ea3 [ 622.321247] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquiring lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.322028] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.322028] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquiring lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.322028] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.322173] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.325227] env[62368]: INFO nova.compute.manager [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Terminating instance [ 622.328517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquiring lock "refresh_cache-d22aff27-739f-4b0d-a0e4-b6316c252c3a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.328517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquired lock "refresh_cache-d22aff27-739f-4b0d-a0e4-b6316c252c3a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.328517] env[62368]: DEBUG nova.network.neutron [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.379615] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.388424] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Successfully created port: 140db14b-33b2-42f7-b7ac-4be9491c1293 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.507151] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04c10536-6858-4da2-9ceb-5f23aedaf159 tempest-ImagesNegativeTestJSON-454980705 tempest-ImagesNegativeTestJSON-454980705-project-member] Lock "1523daf8-033b-48e2-a495-885feb4e0ea3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.481s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.850447] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c118ec-47b6-46f9-b9a8-6446334c68f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.858797] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da56aac-97be-4548-8371-da30eff137fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.862721] env[62368]: DEBUG nova.network.neutron [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.899502] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6bd381-ba36-4c3e-98ca-d154c7c078da {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.907402] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe4cb51-ffd9-4f09-8f71-728b92325be1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.922444] env[62368]: DEBUG nova.compute.provider_tree [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.946382] env[62368]: DEBUG nova.network.neutron [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.014201] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.404554] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.432665] env[62368]: DEBUG nova.scheduler.client.report [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.448033] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.448033] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.448033] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.448532] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.448532] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.448532] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.448705] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.449023] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.449313] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.449598] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.449867] env[62368]: DEBUG nova.virt.hardware [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.450814] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11a35e2-312e-4445-813c-f2c953056b89 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.454445] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Releasing lock "refresh_cache-d22aff27-739f-4b0d-a0e4-b6316c252c3a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.454964] env[62368]: DEBUG nova.compute.manager [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 623.455281] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 623.456411] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c35d64-e07e-4647-a8d5-74d6d43262a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.465146] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 623.467414] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0723415b-d880-4376-9e72-6d7d735ce1e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.470383] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6926cac1-3de2-459b-bb97-b3d7c9d96412 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.490101] env[62368]: DEBUG oslo_vmware.api [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 623.490101] env[62368]: value = "task-1198169" [ 623.490101] env[62368]: _type = "Task" [ 623.490101] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.498582] env[62368]: DEBUG oslo_vmware.api [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.539124] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.600964] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquiring lock "085ebe93-aa24-4626-94fe-241c4297e4db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.601296] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "085ebe93-aa24-4626-94fe-241c4297e4db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.780027] env[62368]: ERROR nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. [ 623.780027] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.780027] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.780027] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.780027] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.780027] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.780027] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.780027] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.780027] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.780027] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 623.780027] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.780027] env[62368]: ERROR nova.compute.manager raise self.value [ 623.780027] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.780027] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.780027] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.780027] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.780466] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.780466] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.780466] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. [ 623.780466] env[62368]: ERROR nova.compute.manager [ 623.780466] env[62368]: Traceback (most recent call last): [ 623.780466] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.780466] env[62368]: listener.cb(fileno) [ 623.780466] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.780466] env[62368]: result = function(*args, **kwargs) [ 623.780466] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.780466] env[62368]: return func(*args, **kwargs) [ 623.780466] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.780466] env[62368]: raise e [ 623.780466] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.780466] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 623.780466] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.780466] env[62368]: created_port_ids = self._update_ports_for_instance( [ 623.780466] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.780466] env[62368]: with excutils.save_and_reraise_exception(): [ 623.780466] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.780466] env[62368]: self.force_reraise() [ 623.780466] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.780466] env[62368]: raise self.value [ 623.780466] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.780466] env[62368]: updated_port = self._update_port( [ 623.780466] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.780466] env[62368]: _ensure_no_port_binding_failure(port) [ 623.780466] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.780466] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.781128] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. [ 623.781128] env[62368]: Removing descriptor: 18 [ 623.781128] env[62368]: ERROR nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Traceback (most recent call last): [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] yield resources [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self.driver.spawn(context, instance, image_meta, [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.781128] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] vm_ref = self.build_virtual_machine(instance, [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] for vif in network_info: [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] return self._sync_wrapper(fn, *args, **kwargs) [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self.wait() [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self[:] = self._gt.wait() [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] return self._exit_event.wait() [ 623.781411] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] result = hub.switch() [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] return self.greenlet.switch() [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] result = function(*args, **kwargs) [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] return func(*args, **kwargs) [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] raise e [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] nwinfo = self.network_api.allocate_for_instance( [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.781773] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] created_port_ids = self._update_ports_for_instance( [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] with excutils.save_and_reraise_exception(): [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self.force_reraise() [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] raise self.value [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] updated_port = self._update_port( [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] _ensure_no_port_binding_failure(port) [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.782111] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] raise exception.PortBindingFailed(port_id=port['id']) [ 623.782409] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] nova.exception.PortBindingFailed: Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. [ 623.782409] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] [ 623.784024] env[62368]: INFO nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Terminating instance [ 623.785717] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.785982] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquired lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.786261] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.885607] env[62368]: DEBUG nova.compute.manager [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Received event network-changed-140db14b-33b2-42f7-b7ac-4be9491c1293 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.885607] env[62368]: DEBUG nova.compute.manager [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Refreshing instance network info cache due to event network-changed-140db14b-33b2-42f7-b7ac-4be9491c1293. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 623.885607] env[62368]: DEBUG oslo_concurrency.lockutils [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] Acquiring lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.942701] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.942701] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 623.943971] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.240s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.944289] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.944544] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 623.944955] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.922s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.946900] env[62368]: INFO nova.compute.claims [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.950026] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6455c17c-6d08-4c53-a90c-3dd0b3a98c8c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.961537] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6653f7-ebdc-4524-9a49-f270d18b1fe2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.978798] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8929b304-cb07-4ac8-8465-03d65757e2e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.989717] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5fa796-3937-4ab6-9c7c-44a18f0865aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.003167] env[62368]: DEBUG oslo_vmware.api [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198169, 'name': PowerOffVM_Task, 'duration_secs': 0.124524} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.028100] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 624.028335] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 624.031141] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181365MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 624.031141] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.031141] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-26ae5423-e6e1-48ea-89dd-0722223cecd8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.053661] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 624.053863] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 624.054045] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Deleting the datastore file [datastore1] d22aff27-739f-4b0d-a0e4-b6316c252c3a {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 624.054292] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2f98efe-0f88-4d3d-963c-1d3329872455 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.060544] env[62368]: DEBUG oslo_vmware.api [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for the task: (returnval){ [ 624.060544] env[62368]: value = "task-1198171" [ 624.060544] env[62368]: _type = "Task" [ 624.060544] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.069035] env[62368]: DEBUG oslo_vmware.api [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198171, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.318877] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.451256] env[62368]: DEBUG nova.compute.utils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.452781] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.453214] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 624.495438] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.538133] env[62368]: DEBUG nova.policy [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72595ab2bbf94ec5ac63a5b5351ae798', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ff43f079c154fc6982d0cd583c6f789', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.573032] env[62368]: DEBUG oslo_vmware.api [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Task: {'id': task-1198171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110136} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.573032] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 624.573032] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 624.573032] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 624.573032] env[62368]: INFO nova.compute.manager [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 624.573322] env[62368]: DEBUG oslo.service.loopingcall [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.573478] env[62368]: DEBUG nova.compute.manager [-] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.573571] env[62368]: DEBUG nova.network.neutron [-] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.604159] env[62368]: DEBUG nova.network.neutron [-] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.957590] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.998074] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Releasing lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.998753] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.999113] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 624.999263] env[62368]: DEBUG oslo_concurrency.lockutils [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] Acquired lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.999455] env[62368]: DEBUG nova.network.neutron [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Refreshing network info cache for port 140db14b-33b2-42f7-b7ac-4be9491c1293 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 625.000906] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d52f303-9356-47db-a562-f1bb82285eda {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.015168] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f927af3-d0cf-43b9-84bb-bf37c8462377 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.034529] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Successfully created port: 29c9a3fc-4255-459d-99c1-edade0cc5220 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.040452] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 45fe27f5-0f20-48ff-ad25-03862ac9bdd9 could not be found. [ 625.040655] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 625.040833] env[62368]: INFO nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 625.041083] env[62368]: DEBUG oslo.service.loopingcall [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.043411] env[62368]: DEBUG nova.compute.manager [-] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.043498] env[62368]: DEBUG nova.network.neutron [-] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.066515] env[62368]: DEBUG nova.network.neutron [-] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.106545] env[62368]: DEBUG nova.network.neutron [-] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.478024] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60165275-160b-4358-a169-70127c9d2802 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.487993] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2eb0dc2-e5c7-4480-ab57-78d57bcff843 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.523962] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96f7934-1916-4377-acb0-c80a27a8f586 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.534201] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d997b06f-3956-418f-9170-a94960e07048 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.563930] env[62368]: DEBUG nova.compute.provider_tree [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 625.563930] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Successfully created port: a62feb05-80db-449b-bf07-430a63af1d82 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.563930] env[62368]: DEBUG nova.network.neutron [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.569130] env[62368]: DEBUG nova.network.neutron [-] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.615858] env[62368]: INFO nova.compute.manager [-] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Took 1.04 seconds to deallocate network for instance. [ 625.671955] env[62368]: DEBUG nova.network.neutron [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.976346] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.020815] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.021230] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.021337] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.021560] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.021703] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.021867] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.022103] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.022279] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.022453] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.022900] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.022900] env[62368]: DEBUG nova.virt.hardware [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.023687] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a232df12-5fbb-48ce-86c7-e14e268b9aa1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.038257] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5209082e-5e11-4ad6-9776-71932ce76917 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.053557] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Successfully created port: 141c20d8-cf59-481e-a2d8-7d4ea5f6be36 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.073274] env[62368]: INFO nova.compute.manager [-] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Took 1.03 seconds to deallocate network for instance. [ 626.076043] env[62368]: DEBUG nova.compute.claims [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.076279] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.089315] env[62368]: ERROR nova.scheduler.client.report [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [req-07e18bab-8779-49b3-8f81-d4c786e30143] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-07e18bab-8779-49b3-8f81-d4c786e30143"}]} [ 626.108531] env[62368]: DEBUG nova.scheduler.client.report [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 626.122649] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.123788] env[62368]: DEBUG nova.scheduler.client.report [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 626.125027] env[62368]: DEBUG nova.compute.provider_tree [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 626.136237] env[62368]: DEBUG nova.scheduler.client.report [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 626.155926] env[62368]: DEBUG nova.scheduler.client.report [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 626.180165] env[62368]: DEBUG oslo_concurrency.lockutils [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] Releasing lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.180266] env[62368]: DEBUG nova.compute.manager [req-76ef2959-83f9-45de-ba8e-b3b29b64242d req-e1444f44-7a58-4cf4-a0ad-c68b98981212 service nova] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Received event network-vif-deleted-140db14b-33b2-42f7-b7ac-4be9491c1293 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.622678] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72da4718-2bed-4b38-b39d-f10d12d5611a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.631286] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b56e003-8927-482e-9fa3-b1f0b1f02740 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.667911] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2553e6-d1f5-4c77-a05b-afd77fb82aa1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.676370] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25cea91-7cf6-48cd-a99d-a69dd4597f6f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.691833] env[62368]: DEBUG nova.compute.provider_tree [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 627.232630] env[62368]: DEBUG nova.scheduler.client.report [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 56 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 627.232923] env[62368]: DEBUG nova.compute.provider_tree [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 56 to 57 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 627.233539] env[62368]: DEBUG nova.compute.provider_tree [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 627.595610] env[62368]: DEBUG nova.compute.manager [req-a9a07d74-3bed-436a-9395-a115d3bfcab3 req-e2dd5445-efd7-4674-837e-b46869904f16 service nova] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Received event network-changed-29c9a3fc-4255-459d-99c1-edade0cc5220 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.595816] env[62368]: DEBUG nova.compute.manager [req-a9a07d74-3bed-436a-9395-a115d3bfcab3 req-e2dd5445-efd7-4674-837e-b46869904f16 service nova] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Refreshing instance network info cache due to event network-changed-29c9a3fc-4255-459d-99c1-edade0cc5220. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 627.596127] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9a07d74-3bed-436a-9395-a115d3bfcab3 req-e2dd5445-efd7-4674-837e-b46869904f16 service nova] Acquiring lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.596312] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9a07d74-3bed-436a-9395-a115d3bfcab3 req-e2dd5445-efd7-4674-837e-b46869904f16 service nova] Acquired lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.596509] env[62368]: DEBUG nova.network.neutron [req-a9a07d74-3bed-436a-9395-a115d3bfcab3 req-e2dd5445-efd7-4674-837e-b46869904f16 service nova] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Refreshing network info cache for port 29c9a3fc-4255-459d-99c1-edade0cc5220 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 627.667193] env[62368]: ERROR nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. [ 627.667193] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.667193] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.667193] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.667193] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.667193] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.667193] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.667193] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.667193] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.667193] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 627.667193] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.667193] env[62368]: ERROR nova.compute.manager raise self.value [ 627.667193] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.667193] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.667193] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.667193] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.667772] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.667772] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.667772] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. [ 627.667772] env[62368]: ERROR nova.compute.manager [ 627.667772] env[62368]: Traceback (most recent call last): [ 627.667772] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.667772] env[62368]: listener.cb(fileno) [ 627.667772] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.667772] env[62368]: result = function(*args, **kwargs) [ 627.667772] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.667772] env[62368]: return func(*args, **kwargs) [ 627.667772] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.667772] env[62368]: raise e [ 627.667772] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.667772] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 627.667772] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.667772] env[62368]: created_port_ids = self._update_ports_for_instance( [ 627.667772] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.667772] env[62368]: with excutils.save_and_reraise_exception(): [ 627.667772] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.667772] env[62368]: self.force_reraise() [ 627.667772] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.667772] env[62368]: raise self.value [ 627.667772] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.667772] env[62368]: updated_port = self._update_port( [ 627.667772] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.667772] env[62368]: _ensure_no_port_binding_failure(port) [ 627.667772] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.667772] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.668980] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. [ 627.668980] env[62368]: Removing descriptor: 15 [ 627.668980] env[62368]: ERROR nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Traceback (most recent call last): [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] yield resources [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self.driver.spawn(context, instance, image_meta, [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.668980] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] vm_ref = self.build_virtual_machine(instance, [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] for vif in network_info: [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] return self._sync_wrapper(fn, *args, **kwargs) [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self.wait() [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self[:] = self._gt.wait() [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] return self._exit_event.wait() [ 627.669377] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] result = hub.switch() [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] return self.greenlet.switch() [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] result = function(*args, **kwargs) [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] return func(*args, **kwargs) [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] raise e [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] nwinfo = self.network_api.allocate_for_instance( [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.669873] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] created_port_ids = self._update_ports_for_instance( [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] with excutils.save_and_reraise_exception(): [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self.force_reraise() [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] raise self.value [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] updated_port = self._update_port( [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] _ensure_no_port_binding_failure(port) [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.670269] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] raise exception.PortBindingFailed(port_id=port['id']) [ 627.670656] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] nova.exception.PortBindingFailed: Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. [ 627.670656] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] [ 627.670656] env[62368]: INFO nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Terminating instance [ 627.670656] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.741408] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.796s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.741941] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 627.745255] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.898s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.749325] env[62368]: INFO nova.compute.claims [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.152557] env[62368]: DEBUG nova.network.neutron [req-a9a07d74-3bed-436a-9395-a115d3bfcab3 req-e2dd5445-efd7-4674-837e-b46869904f16 service nova] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.250877] env[62368]: DEBUG nova.compute.utils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.256369] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.258875] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 628.264996] env[62368]: DEBUG nova.network.neutron [req-a9a07d74-3bed-436a-9395-a115d3bfcab3 req-e2dd5445-efd7-4674-837e-b46869904f16 service nova] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.341856] env[62368]: DEBUG nova.policy [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81791f8011614349806f125ed9c392da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6bb9d26749748e4adb3c63225bb9bde', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.515615] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.515839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.757189] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 628.773121] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9a07d74-3bed-436a-9395-a115d3bfcab3 req-e2dd5445-efd7-4674-837e-b46869904f16 service nova] Releasing lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.773121] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquired lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.773121] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.805409] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Successfully created port: d7d96866-a2f9-44a6-84e1-6fa2761c7162 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.229879] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3f6c01-50b6-44b2-bbd2-027667a274e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.244296] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141467bb-df9b-4b3e-b567-e9624f984d1d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.285374] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7608c1-b7e6-4024-9c36-501e3a849062 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.292915] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6112c5e0-32c6-4be7-8774-3872566cd8bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.306263] env[62368]: DEBUG nova.compute.provider_tree [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.336020] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.463202] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.724127] env[62368]: DEBUG nova.compute.manager [req-488b25b7-76f3-4883-aad3-13968f8253eb req-553f6969-d505-4c1f-9290-fe235ebb94f5 service nova] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Received event network-vif-deleted-29c9a3fc-4255-459d-99c1-edade0cc5220 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.786291] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.809101] env[62368]: DEBUG nova.scheduler.client.report [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.815022] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.815022] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.816558] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.816558] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.816678] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.816743] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.817754] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.817754] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.817754] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.817754] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.818218] env[62368]: DEBUG nova.virt.hardware [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.819086] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26055001-9116-401d-8ee0-16e90306029b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.828214] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6f9cc8-b278-4369-8fa1-9600881f9efb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.965620] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Releasing lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.966350] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 629.966590] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 629.967572] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d78b14e3-7e93-425d-9c21-9965cf2debdd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.980633] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb2b440-14f2-43b8-8a54-b2d3cfa913aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.011778] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bebed5b9-d067-4cf2-a203-48763685ecb6 could not be found. [ 630.012223] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 630.013062] env[62368]: INFO nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 630.014087] env[62368]: DEBUG oslo.service.loopingcall [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.014087] env[62368]: DEBUG nova.compute.manager [-] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.014087] env[62368]: DEBUG nova.network.neutron [-] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 630.049567] env[62368]: DEBUG nova.network.neutron [-] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.146217] env[62368]: DEBUG nova.compute.manager [req-0608411e-84a9-4d1a-aad5-b2fbabf3585d req-2af00d41-ff8a-490f-8a75-9caebcefaae5 service nova] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Received event network-changed-d7d96866-a2f9-44a6-84e1-6fa2761c7162 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.146217] env[62368]: DEBUG nova.compute.manager [req-0608411e-84a9-4d1a-aad5-b2fbabf3585d req-2af00d41-ff8a-490f-8a75-9caebcefaae5 service nova] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Refreshing instance network info cache due to event network-changed-d7d96866-a2f9-44a6-84e1-6fa2761c7162. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 630.146217] env[62368]: DEBUG oslo_concurrency.lockutils [req-0608411e-84a9-4d1a-aad5-b2fbabf3585d req-2af00d41-ff8a-490f-8a75-9caebcefaae5 service nova] Acquiring lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.146217] env[62368]: DEBUG oslo_concurrency.lockutils [req-0608411e-84a9-4d1a-aad5-b2fbabf3585d req-2af00d41-ff8a-490f-8a75-9caebcefaae5 service nova] Acquired lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.146217] env[62368]: DEBUG nova.network.neutron [req-0608411e-84a9-4d1a-aad5-b2fbabf3585d req-2af00d41-ff8a-490f-8a75-9caebcefaae5 service nova] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Refreshing network info cache for port d7d96866-a2f9-44a6-84e1-6fa2761c7162 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 630.289707] env[62368]: ERROR nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. [ 630.289707] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.289707] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.289707] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.289707] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.289707] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.289707] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.289707] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.289707] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.289707] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 630.289707] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.289707] env[62368]: ERROR nova.compute.manager raise self.value [ 630.289707] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.289707] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.289707] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.289707] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.290342] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.290342] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.290342] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. [ 630.290342] env[62368]: ERROR nova.compute.manager [ 630.290342] env[62368]: Traceback (most recent call last): [ 630.290342] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.290342] env[62368]: listener.cb(fileno) [ 630.290342] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.290342] env[62368]: result = function(*args, **kwargs) [ 630.290342] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.290342] env[62368]: return func(*args, **kwargs) [ 630.290342] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.290342] env[62368]: raise e [ 630.290342] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.290342] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 630.290342] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.290342] env[62368]: created_port_ids = self._update_ports_for_instance( [ 630.290342] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.290342] env[62368]: with excutils.save_and_reraise_exception(): [ 630.290342] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.290342] env[62368]: self.force_reraise() [ 630.290342] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.290342] env[62368]: raise self.value [ 630.290342] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.290342] env[62368]: updated_port = self._update_port( [ 630.290342] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.290342] env[62368]: _ensure_no_port_binding_failure(port) [ 630.290342] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.290342] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.291088] env[62368]: nova.exception.PortBindingFailed: Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. [ 630.291088] env[62368]: Removing descriptor: 18 [ 630.291088] env[62368]: ERROR nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] Traceback (most recent call last): [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] yield resources [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self.driver.spawn(context, instance, image_meta, [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.291088] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] vm_ref = self.build_virtual_machine(instance, [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] for vif in network_info: [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] return self._sync_wrapper(fn, *args, **kwargs) [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self.wait() [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self[:] = self._gt.wait() [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] return self._exit_event.wait() [ 630.291434] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] result = hub.switch() [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] return self.greenlet.switch() [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] result = function(*args, **kwargs) [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] return func(*args, **kwargs) [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] raise e [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] nwinfo = self.network_api.allocate_for_instance( [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.291759] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] created_port_ids = self._update_ports_for_instance( [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] with excutils.save_and_reraise_exception(): [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self.force_reraise() [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] raise self.value [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] updated_port = self._update_port( [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] _ensure_no_port_binding_failure(port) [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.292148] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] raise exception.PortBindingFailed(port_id=port['id']) [ 630.292480] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] nova.exception.PortBindingFailed: Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. [ 630.292480] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] [ 630.292480] env[62368]: INFO nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Terminating instance [ 630.293975] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Acquiring lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.326795] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.582s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.327332] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.331568] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.935s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.331568] env[62368]: INFO nova.compute.claims [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.669747] env[62368]: DEBUG nova.network.neutron [req-0608411e-84a9-4d1a-aad5-b2fbabf3585d req-2af00d41-ff8a-490f-8a75-9caebcefaae5 service nova] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.704220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.704430] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.765386] env[62368]: DEBUG nova.network.neutron [req-0608411e-84a9-4d1a-aad5-b2fbabf3585d req-2af00d41-ff8a-490f-8a75-9caebcefaae5 service nova] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.842021] env[62368]: DEBUG nova.compute.utils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.842911] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 630.843111] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 630.894580] env[62368]: DEBUG nova.policy [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f53480e7585471fbf084ba6603d4ab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bdba0baa56ae465ca0045f41da157304', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.024947] env[62368]: DEBUG nova.network.neutron [-] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.195250] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Successfully created port: 6703e135-9976-4bbf-9fad-b4cec2b0f20b {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.267639] env[62368]: DEBUG oslo_concurrency.lockutils [req-0608411e-84a9-4d1a-aad5-b2fbabf3585d req-2af00d41-ff8a-490f-8a75-9caebcefaae5 service nova] Releasing lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.268054] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Acquired lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.268239] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 631.347966] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 631.529136] env[62368]: INFO nova.compute.manager [-] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Took 1.52 seconds to deallocate network for instance. [ 631.532262] env[62368]: DEBUG nova.compute.claims [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 631.532262] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.800525] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.859109] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0d0071-2895-41c9-a433-9b69d19573fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.867304] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab008fa-4540-4509-a331-5ff884de89ce {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.899430] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.901429] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee873e6-8e84-4bb1-af3a-18c8d4503f2a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.909391] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22889f0-70d9-4769-88f6-83f00dfb1cd1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.923354] env[62368]: DEBUG nova.compute.provider_tree [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.001671] env[62368]: DEBUG nova.compute.manager [req-b9c6a0df-a3a6-47fb-b987-de44d949e103 req-9f98704b-4711-4092-a7e1-b9fc538a6e19 service nova] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Received event network-changed-6703e135-9976-4bbf-9fad-b4cec2b0f20b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.001911] env[62368]: DEBUG nova.compute.manager [req-b9c6a0df-a3a6-47fb-b987-de44d949e103 req-9f98704b-4711-4092-a7e1-b9fc538a6e19 service nova] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Refreshing instance network info cache due to event network-changed-6703e135-9976-4bbf-9fad-b4cec2b0f20b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 632.002173] env[62368]: DEBUG oslo_concurrency.lockutils [req-b9c6a0df-a3a6-47fb-b987-de44d949e103 req-9f98704b-4711-4092-a7e1-b9fc538a6e19 service nova] Acquiring lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.002377] env[62368]: DEBUG oslo_concurrency.lockutils [req-b9c6a0df-a3a6-47fb-b987-de44d949e103 req-9f98704b-4711-4092-a7e1-b9fc538a6e19 service nova] Acquired lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.002608] env[62368]: DEBUG nova.network.neutron [req-b9c6a0df-a3a6-47fb-b987-de44d949e103 req-9f98704b-4711-4092-a7e1-b9fc538a6e19 service nova] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Refreshing network info cache for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 632.191345] env[62368]: ERROR nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. [ 632.191345] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.191345] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.191345] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.191345] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.191345] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.191345] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.191345] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.191345] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.191345] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 632.191345] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.191345] env[62368]: ERROR nova.compute.manager raise self.value [ 632.191345] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.191345] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.191345] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.191345] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.191813] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.191813] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.191813] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. [ 632.191813] env[62368]: ERROR nova.compute.manager [ 632.191813] env[62368]: Traceback (most recent call last): [ 632.191813] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.191813] env[62368]: listener.cb(fileno) [ 632.191813] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.191813] env[62368]: result = function(*args, **kwargs) [ 632.191813] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.191813] env[62368]: return func(*args, **kwargs) [ 632.191813] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.191813] env[62368]: raise e [ 632.191813] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.191813] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 632.191813] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.191813] env[62368]: created_port_ids = self._update_ports_for_instance( [ 632.191813] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.191813] env[62368]: with excutils.save_and_reraise_exception(): [ 632.191813] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.191813] env[62368]: self.force_reraise() [ 632.191813] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.191813] env[62368]: raise self.value [ 632.191813] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.191813] env[62368]: updated_port = self._update_port( [ 632.191813] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.191813] env[62368]: _ensure_no_port_binding_failure(port) [ 632.191813] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.191813] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.192609] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. [ 632.192609] env[62368]: Removing descriptor: 15 [ 632.218829] env[62368]: DEBUG nova.compute.manager [req-0217ca0f-b656-46c4-87c1-fedf2f00522e req-36843462-928d-4f32-8df4-1d6d06be9826 service nova] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Received event network-vif-deleted-d7d96866-a2f9-44a6-84e1-6fa2761c7162 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.359537] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.385459] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.385708] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.385862] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.386060] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.386221] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.386366] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.386572] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.386728] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.386892] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.387064] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.387243] env[62368]: DEBUG nova.virt.hardware [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.388104] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b5e0d7-11d5-4794-9bc0-9f4ec0ba67d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.396232] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494acd16-cec7-4ea1-a8a3-af64b3e024fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.410629] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Releasing lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.411017] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.411210] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 632.412072] env[62368]: ERROR nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Traceback (most recent call last): [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] yield resources [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self.driver.spawn(context, instance, image_meta, [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] vm_ref = self.build_virtual_machine(instance, [ 632.412072] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] for vif in network_info: [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] return self._sync_wrapper(fn, *args, **kwargs) [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self.wait() [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self[:] = self._gt.wait() [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] return self._exit_event.wait() [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 632.412428] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] current.throw(*self._exc) [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] result = function(*args, **kwargs) [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] return func(*args, **kwargs) [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] raise e [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] nwinfo = self.network_api.allocate_for_instance( [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] created_port_ids = self._update_ports_for_instance( [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] with excutils.save_and_reraise_exception(): [ 632.412962] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self.force_reraise() [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] raise self.value [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] updated_port = self._update_port( [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] _ensure_no_port_binding_failure(port) [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] raise exception.PortBindingFailed(port_id=port['id']) [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] nova.exception.PortBindingFailed: Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. [ 632.413447] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] [ 632.413447] env[62368]: INFO nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Terminating instance [ 632.413873] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95581f25-13f0-4f58-aee7-00063fcb3f40 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.415781] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.422094] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-465b3838-55a6-4c74-bdbd-bda6fc9f6e4a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.432504] env[62368]: DEBUG nova.scheduler.client.report [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.445882] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 63c88ac7-9700-4476-9039-feb034de659d could not be found. [ 632.446104] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 632.446286] env[62368]: INFO nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 632.446523] env[62368]: DEBUG oslo.service.loopingcall [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.447282] env[62368]: DEBUG nova.compute.manager [-] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.447376] env[62368]: DEBUG nova.network.neutron [-] [instance: 63c88ac7-9700-4476-9039-feb034de659d] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 632.461523] env[62368]: DEBUG nova.network.neutron [-] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.522745] env[62368]: DEBUG nova.network.neutron [req-b9c6a0df-a3a6-47fb-b987-de44d949e103 req-9f98704b-4711-4092-a7e1-b9fc538a6e19 service nova] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.581286] env[62368]: DEBUG nova.network.neutron [req-b9c6a0df-a3a6-47fb-b987-de44d949e103 req-9f98704b-4711-4092-a7e1-b9fc538a6e19 service nova] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.937233] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.607s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.937764] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 632.940294] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.483s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.941769] env[62368]: INFO nova.compute.claims [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.963616] env[62368]: DEBUG nova.network.neutron [-] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.084431] env[62368]: DEBUG oslo_concurrency.lockutils [req-b9c6a0df-a3a6-47fb-b987-de44d949e103 req-9f98704b-4711-4092-a7e1-b9fc538a6e19 service nova] Releasing lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.084888] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquired lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.085254] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.445922] env[62368]: DEBUG nova.compute.utils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.449863] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.450086] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 633.466036] env[62368]: INFO nova.compute.manager [-] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Took 1.02 seconds to deallocate network for instance. [ 633.468242] env[62368]: DEBUG nova.compute.claims [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.468756] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.509950] env[62368]: DEBUG nova.policy [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5944b43a34d64afabf8e23bb70fdc484', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9f7fd8890224444b82ae1ef9c12bdc9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.602666] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.659344] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.799399] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Successfully created port: cfc267e0-8cab-4795-8a71-ca70949ca14f {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.951304] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.028333] env[62368]: DEBUG nova.compute.manager [req-d4b0a1a1-a1da-4b7d-a813-56f5130eaa0e req-cf04f6f2-b64e-4d74-a9c5-3caa435fa961 service nova] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Received event network-vif-deleted-6703e135-9976-4bbf-9fad-b4cec2b0f20b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.162093] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Releasing lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.162786] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 634.163114] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 634.163808] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0a0addd-5936-4e16-b278-f87023058536 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.173516] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d88bfc-d997-47a7-8b3f-c8eaa79f1b7d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.197834] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b0bcd561-35cd-475f-975e-e78456a87fac could not be found. [ 634.200724] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 634.200724] env[62368]: INFO nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Took 0.04 seconds to destroy the instance on the hypervisor. [ 634.200724] env[62368]: DEBUG oslo.service.loopingcall [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.200724] env[62368]: DEBUG nova.compute.manager [-] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.200724] env[62368]: DEBUG nova.network.neutron [-] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.216108] env[62368]: DEBUG nova.network.neutron [-] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.379378] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2613251c-8ecf-44f2-95fb-d37048d3a7cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.388054] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7aad34a-e70c-4870-b63e-04e7f19d7a5a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.418333] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a497d5a9-4891-4181-9b71-c3e5a4a9004d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.425204] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7931fa-cab1-411d-bc13-ee2fad186851 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.438307] env[62368]: DEBUG nova.compute.provider_tree [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.722330] env[62368]: DEBUG nova.network.neutron [-] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.895611] env[62368]: ERROR nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. [ 634.895611] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.895611] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.895611] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.895611] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.895611] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.895611] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.895611] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.895611] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.895611] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 634.895611] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.895611] env[62368]: ERROR nova.compute.manager raise self.value [ 634.895611] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.895611] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.895611] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.895611] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.896366] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.896366] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.896366] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. [ 634.896366] env[62368]: ERROR nova.compute.manager [ 634.896366] env[62368]: Traceback (most recent call last): [ 634.896366] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.896366] env[62368]: listener.cb(fileno) [ 634.896366] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.896366] env[62368]: result = function(*args, **kwargs) [ 634.896366] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.896366] env[62368]: return func(*args, **kwargs) [ 634.896366] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.896366] env[62368]: raise e [ 634.896366] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.896366] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 634.896366] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.896366] env[62368]: created_port_ids = self._update_ports_for_instance( [ 634.896366] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.896366] env[62368]: with excutils.save_and_reraise_exception(): [ 634.896366] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.896366] env[62368]: self.force_reraise() [ 634.896366] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.896366] env[62368]: raise self.value [ 634.896366] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.896366] env[62368]: updated_port = self._update_port( [ 634.896366] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.896366] env[62368]: _ensure_no_port_binding_failure(port) [ 634.896366] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.896366] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.897727] env[62368]: nova.exception.PortBindingFailed: Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. [ 634.897727] env[62368]: Removing descriptor: 15 [ 634.941996] env[62368]: DEBUG nova.scheduler.client.report [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.963910] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 634.990036] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.990279] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.990429] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.990683] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.990849] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.991058] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.991222] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.991399] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.991565] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.991755] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.991952] env[62368]: DEBUG nova.virt.hardware [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.992847] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46cf4bc-cd12-464c-9f2c-7888b3020498 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.003182] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ba867b-eb1a-439d-97ec-7f29630ceb87 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.017073] env[62368]: ERROR nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Traceback (most recent call last): [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] yield resources [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self.driver.spawn(context, instance, image_meta, [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] vm_ref = self.build_virtual_machine(instance, [ 635.017073] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] for vif in network_info: [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] return self._sync_wrapper(fn, *args, **kwargs) [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self.wait() [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self[:] = self._gt.wait() [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] return self._exit_event.wait() [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 635.017750] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] current.throw(*self._exc) [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] result = function(*args, **kwargs) [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] return func(*args, **kwargs) [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] raise e [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] nwinfo = self.network_api.allocate_for_instance( [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] created_port_ids = self._update_ports_for_instance( [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] with excutils.save_and_reraise_exception(): [ 635.018170] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self.force_reraise() [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] raise self.value [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] updated_port = self._update_port( [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] _ensure_no_port_binding_failure(port) [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] raise exception.PortBindingFailed(port_id=port['id']) [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] nova.exception.PortBindingFailed: Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. [ 635.018506] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] [ 635.018506] env[62368]: INFO nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Terminating instance [ 635.019767] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Acquiring lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.019767] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Acquired lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.019922] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.225606] env[62368]: INFO nova.compute.manager [-] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Took 1.03 seconds to deallocate network for instance. [ 635.228330] env[62368]: DEBUG nova.compute.claims [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.228330] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.447052] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.447512] env[62368]: DEBUG nova.compute.manager [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.450539] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.538s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.451990] env[62368]: INFO nova.compute.claims [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.541077] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.643933] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.958012] env[62368]: DEBUG nova.compute.utils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 635.962032] env[62368]: DEBUG nova.compute.manager [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Not allocating networking since 'none' was specified. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 636.052219] env[62368]: DEBUG nova.compute.manager [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Received event network-changed-cfc267e0-8cab-4795-8a71-ca70949ca14f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.052436] env[62368]: DEBUG nova.compute.manager [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Refreshing instance network info cache due to event network-changed-cfc267e0-8cab-4795-8a71-ca70949ca14f. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 636.052650] env[62368]: DEBUG oslo_concurrency.lockutils [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] Acquiring lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.147564] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Releasing lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.148028] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.148228] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 636.148531] env[62368]: DEBUG oslo_concurrency.lockutils [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] Acquired lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.148697] env[62368]: DEBUG nova.network.neutron [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Refreshing network info cache for port cfc267e0-8cab-4795-8a71-ca70949ca14f {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.149815] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bebcd3c-08e7-4a7e-b348-160090b56a35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.158867] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c89661e-d5e6-4c46-83a0-b546a41b980d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.181064] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bcecb4c6-6675-403d-9a3d-b96156b1d56a could not be found. [ 636.181188] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 636.181343] env[62368]: INFO nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 636.181586] env[62368]: DEBUG oslo.service.loopingcall [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.181803] env[62368]: DEBUG nova.compute.manager [-] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.181903] env[62368]: DEBUG nova.network.neutron [-] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.203954] env[62368]: DEBUG nova.network.neutron [-] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.463719] env[62368]: DEBUG nova.compute.manager [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.671097] env[62368]: DEBUG nova.network.neutron [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.706148] env[62368]: DEBUG nova.network.neutron [-] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.755577] env[62368]: DEBUG nova.network.neutron [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.831341] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a6cee8-9542-472b-be5c-d0cc54c2298d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.839139] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa5202c-da3e-41d4-b5d8-ee393dabb9b2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.875437] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3122888-0d96-4e75-9596-20917ab0df85 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.882990] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18deb1c-2adb-44d7-929f-330faa3346ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.899015] env[62368]: DEBUG nova.compute.provider_tree [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.208539] env[62368]: INFO nova.compute.manager [-] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Took 1.03 seconds to deallocate network for instance. [ 637.210926] env[62368]: DEBUG nova.compute.claims [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.211115] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.257920] env[62368]: DEBUG oslo_concurrency.lockutils [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] Releasing lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.258232] env[62368]: DEBUG nova.compute.manager [req-7ad3077e-dcc0-4e99-a9cc-7d66962b2086 req-a2a8b1ee-e250-4e3e-a537-fbda972f64cb service nova] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Received event network-vif-deleted-cfc267e0-8cab-4795-8a71-ca70949ca14f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.401270] env[62368]: DEBUG nova.scheduler.client.report [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.475211] env[62368]: DEBUG nova.compute.manager [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.499737] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.499975] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.500159] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.500350] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.500494] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.500633] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.500839] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.500994] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.501175] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.501338] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.501505] env[62368]: DEBUG nova.virt.hardware [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.502428] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e7e921-bf0f-4d7a-8509-10e863e633c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.510130] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8880af3c-4385-4b63-9059-34998cfa150c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.525719] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 637.529537] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Creating folder: Project (c8c646b77fe44361acaa32bf84c66185). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 637.530987] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63d733b3-9fc0-4e77-99a7-6bc18bf20405 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.539544] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Created folder: Project (c8c646b77fe44361acaa32bf84c66185) in parent group-v259706. [ 637.539724] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Creating folder: Instances. Parent ref: group-v259719. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 637.539939] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-acc0ceb0-352f-455b-89ef-8a5b1707449c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.548032] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Created folder: Instances in parent group-v259719. [ 637.548258] env[62368]: DEBUG oslo.service.loopingcall [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.548437] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 637.548622] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8e6253b-ae59-41ac-8f39-dc164219e90a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.564667] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 637.564667] env[62368]: value = "task-1198174" [ 637.564667] env[62368]: _type = "Task" [ 637.564667] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.571367] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198174, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.906640] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.907402] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.909958] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.003s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.911521] env[62368]: INFO nova.compute.claims [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.074494] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198174, 'name': CreateVM_Task, 'duration_secs': 0.243989} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.074668] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 638.075121] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.075251] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.075602] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 638.076114] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-140405a9-700c-449c-b461-bde72ca82f6b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.080631] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 638.080631] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5281a56c-4547-043b-3cb2-38a87c38a4b3" [ 638.080631] env[62368]: _type = "Task" [ 638.080631] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.088127] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5281a56c-4547-043b-3cb2-38a87c38a4b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.416211] env[62368]: DEBUG nova.compute.utils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.419661] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 638.419836] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 638.456608] env[62368]: DEBUG nova.policy [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68c2209271e043b698fd3794955f40bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcecae6c09124e9fa4dcd5d63833b0bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 638.593855] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5281a56c-4547-043b-3cb2-38a87c38a4b3, 'name': SearchDatastore_Task, 'duration_secs': 0.008476} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.594220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.594471] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 638.594698] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.594845] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.595035] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 638.595305] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a36c844-838f-4270-8ef3-6e25d89bdc26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.603469] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 638.603751] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 638.604831] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2bf1a92-2600-4939-b969-9e82deae47d4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.611445] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 638.611445] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52212bba-f3f9-49b4-3857-0fdd44841b09" [ 638.611445] env[62368]: _type = "Task" [ 638.611445] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.622422] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52212bba-f3f9-49b4-3857-0fdd44841b09, 'name': SearchDatastore_Task, 'duration_secs': 0.008771} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.622700] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39a24dcd-ac9e-4d20-8bb8-ad3eea5eb659 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.627276] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 638.627276] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52922a22-e8ab-e732-ab2d-54474998c313" [ 638.627276] env[62368]: _type = "Task" [ 638.627276] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.634758] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52922a22-e8ab-e732-ab2d-54474998c313, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.769229] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Successfully created port: abe3f272-595d-431f-9af7-e237e3f8164c {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.920403] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 639.141385] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52922a22-e8ab-e732-ab2d-54474998c313, 'name': SearchDatastore_Task, 'duration_secs': 0.013778} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.141683] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.141991] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24/4fde6065-41a7-4a95-96df-e442762dad24.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 639.142311] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0701dd2c-ede8-4b35-823a-ab4dd86bc9af {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.149367] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 639.149367] env[62368]: value = "task-1198175" [ 639.149367] env[62368]: _type = "Task" [ 639.149367] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.160529] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.400235] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd597215-9a46-4263-b9e4-f613f81f6ad6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.408742] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044ac1e2-e43c-4f09-9663-b2f252899b77 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.448220] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13074d1-1aa8-4e28-ac9e-0d4aec56e43c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.464769] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33775c4-1042-41ff-8eea-8fedcf20aeca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.487515] env[62368]: DEBUG nova.compute.provider_tree [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.666424] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198175, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.684638] env[62368]: DEBUG nova.compute.manager [req-21c7c9be-d42a-41da-a1f7-1bfc865b443b req-c62375e4-d88a-4579-a479-383c0824306d service nova] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Received event network-changed-abe3f272-595d-431f-9af7-e237e3f8164c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.684835] env[62368]: DEBUG nova.compute.manager [req-21c7c9be-d42a-41da-a1f7-1bfc865b443b req-c62375e4-d88a-4579-a479-383c0824306d service nova] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Refreshing instance network info cache due to event network-changed-abe3f272-595d-431f-9af7-e237e3f8164c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 639.685061] env[62368]: DEBUG oslo_concurrency.lockutils [req-21c7c9be-d42a-41da-a1f7-1bfc865b443b req-c62375e4-d88a-4579-a479-383c0824306d service nova] Acquiring lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.685205] env[62368]: DEBUG oslo_concurrency.lockutils [req-21c7c9be-d42a-41da-a1f7-1bfc865b443b req-c62375e4-d88a-4579-a479-383c0824306d service nova] Acquired lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.685624] env[62368]: DEBUG nova.network.neutron [req-21c7c9be-d42a-41da-a1f7-1bfc865b443b req-c62375e4-d88a-4579-a479-383c0824306d service nova] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Refreshing network info cache for port abe3f272-595d-431f-9af7-e237e3f8164c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 639.889739] env[62368]: ERROR nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. [ 639.889739] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 639.889739] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.889739] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 639.889739] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.889739] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 639.889739] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.889739] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 639.889739] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.889739] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 639.889739] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.889739] env[62368]: ERROR nova.compute.manager raise self.value [ 639.889739] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.889739] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 639.889739] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.889739] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 639.890490] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.890490] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 639.890490] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. [ 639.890490] env[62368]: ERROR nova.compute.manager [ 639.890490] env[62368]: Traceback (most recent call last): [ 639.890490] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 639.890490] env[62368]: listener.cb(fileno) [ 639.890490] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.890490] env[62368]: result = function(*args, **kwargs) [ 639.890490] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.890490] env[62368]: return func(*args, **kwargs) [ 639.890490] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.890490] env[62368]: raise e [ 639.890490] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.890490] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 639.890490] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.890490] env[62368]: created_port_ids = self._update_ports_for_instance( [ 639.890490] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.890490] env[62368]: with excutils.save_and_reraise_exception(): [ 639.890490] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.890490] env[62368]: self.force_reraise() [ 639.890490] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.890490] env[62368]: raise self.value [ 639.890490] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.890490] env[62368]: updated_port = self._update_port( [ 639.890490] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.890490] env[62368]: _ensure_no_port_binding_failure(port) [ 639.890490] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.890490] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 639.891686] env[62368]: nova.exception.PortBindingFailed: Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. [ 639.891686] env[62368]: Removing descriptor: 15 [ 639.953398] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.978557] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.978686] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.978793] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.978974] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.979181] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.979331] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.979542] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.979712] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.979874] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.980070] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.980276] env[62368]: DEBUG nova.virt.hardware [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.981166] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c90d67-59af-4461-a519-852559b09e90 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.989436] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e19517-6f96-4602-8eb8-3334266644b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.993802] env[62368]: DEBUG nova.scheduler.client.report [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.011816] env[62368]: ERROR nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Traceback (most recent call last): [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] yield resources [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self.driver.spawn(context, instance, image_meta, [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] vm_ref = self.build_virtual_machine(instance, [ 640.011816] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] for vif in network_info: [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] return self._sync_wrapper(fn, *args, **kwargs) [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self.wait() [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self[:] = self._gt.wait() [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] return self._exit_event.wait() [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 640.012352] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] current.throw(*self._exc) [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] result = function(*args, **kwargs) [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] return func(*args, **kwargs) [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] raise e [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] nwinfo = self.network_api.allocate_for_instance( [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] created_port_ids = self._update_ports_for_instance( [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] with excutils.save_and_reraise_exception(): [ 640.012771] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self.force_reraise() [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] raise self.value [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] updated_port = self._update_port( [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] _ensure_no_port_binding_failure(port) [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] raise exception.PortBindingFailed(port_id=port['id']) [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] nova.exception.PortBindingFailed: Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. [ 640.013211] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] [ 640.013211] env[62368]: INFO nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Terminating instance [ 640.013853] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.164132] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.593647} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.164132] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24/4fde6065-41a7-4a95-96df-e442762dad24.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 640.164132] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 640.164415] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75a0ea53-0442-4f0c-929f-e7e07a0de862 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.170500] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 640.170500] env[62368]: value = "task-1198176" [ 640.170500] env[62368]: _type = "Task" [ 640.170500] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.178590] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198176, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.205746] env[62368]: DEBUG nova.network.neutron [req-21c7c9be-d42a-41da-a1f7-1bfc865b443b req-c62375e4-d88a-4579-a479-383c0824306d service nova] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.297549] env[62368]: DEBUG nova.network.neutron [req-21c7c9be-d42a-41da-a1f7-1bfc865b443b req-c62375e4-d88a-4579-a479-383c0824306d service nova] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.499519] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.500189] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 640.503347] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.705s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.680275] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198176, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083958} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.680534] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 640.681478] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9cde2e-a153-4f1a-aa29-615f9712b0e3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.701355] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24/4fde6065-41a7-4a95-96df-e442762dad24.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 640.701575] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b23fe279-e4a4-4cd1-bd15-47175e24e9c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.720718] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 640.720718] env[62368]: value = "task-1198177" [ 640.720718] env[62368]: _type = "Task" [ 640.720718] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.728241] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198177, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.800708] env[62368]: DEBUG oslo_concurrency.lockutils [req-21c7c9be-d42a-41da-a1f7-1bfc865b443b req-c62375e4-d88a-4579-a479-383c0824306d service nova] Releasing lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.801138] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.801328] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.008610] env[62368]: DEBUG nova.compute.utils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 641.013108] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 641.013260] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 641.051328] env[62368]: DEBUG nova.policy [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e17ec74ebf8d4c12accd841e4503b0df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23e257a2707d4c5cb3062081be3c0546', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 641.233159] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198177, 'name': ReconfigVM_Task, 'duration_secs': 0.272927} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.233442] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24/4fde6065-41a7-4a95-96df-e442762dad24.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 641.234071] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ee814b0-da02-4b2e-a850-e1ac0fa99a7d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.242613] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 641.242613] env[62368]: value = "task-1198178" [ 641.242613] env[62368]: _type = "Task" [ 641.242613] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.250222] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198178, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.322223] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.359173] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Successfully created port: c6adff2e-9829-4e58-85bd-ebadf3be3437 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.399495] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.410862] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52218611-e810-4c8d-9cb4-1b7bebea6f05 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.421135] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffb6820-5d16-4201-9b04-293bc4f2978c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.450980] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc37018-f9b1-4548-b729-c30bdc0de303 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.458576] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657307f8-2dfb-43fd-968b-383d7122e5cd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.473187] env[62368]: DEBUG nova.compute.provider_tree [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 641.515502] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.712699] env[62368]: DEBUG nova.compute.manager [req-9d0cc8df-7c7d-4416-b1be-6eaa819549a3 req-009c9b20-e240-4dbd-a88b-84b816e45ee0 service nova] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Received event network-vif-deleted-abe3f272-595d-431f-9af7-e237e3f8164c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.754605] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198178, 'name': Rename_Task, 'duration_secs': 0.132198} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.754884] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 641.755130] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd134ac4-4b0f-452c-a7f7-4e3d2141ff5e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.761147] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 641.761147] env[62368]: value = "task-1198179" [ 641.761147] env[62368]: _type = "Task" [ 641.761147] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.770308] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198179, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.902703] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.903142] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 641.903334] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 641.903712] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-420fbaa1-bffe-4548-bf1d-ed7ddf13b1dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.913132] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cac898-bbaa-4194-ae01-885ee287b8f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.937897] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9f55aeb-273b-4dae-b871-cfb1572e6535 could not be found. [ 641.938145] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 641.938331] env[62368]: INFO nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Took 0.04 seconds to destroy the instance on the hypervisor. [ 641.938890] env[62368]: DEBUG oslo.service.loopingcall [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.938890] env[62368]: DEBUG nova.compute.manager [-] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.938890] env[62368]: DEBUG nova.network.neutron [-] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 641.957282] env[62368]: DEBUG nova.network.neutron [-] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.993556] env[62368]: ERROR nova.scheduler.client.report [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [req-715226da-3c0b-496c-b184-53b785a7413d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-715226da-3c0b-496c-b184-53b785a7413d"}]}: nova.exception.PortBindingFailed: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. [ 642.016438] env[62368]: DEBUG nova.scheduler.client.report [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 642.034514] env[62368]: DEBUG nova.scheduler.client.report [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 642.034866] env[62368]: DEBUG nova.compute.provider_tree [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 642.054396] env[62368]: DEBUG nova.scheduler.client.report [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 642.073963] env[62368]: DEBUG nova.scheduler.client.report [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 642.272901] env[62368]: DEBUG oslo_vmware.api [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198179, 'name': PowerOnVM_Task, 'duration_secs': 0.424911} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.273228] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 642.273410] env[62368]: INFO nova.compute.manager [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Took 4.80 seconds to spawn the instance on the hypervisor. [ 642.273619] env[62368]: DEBUG nova.compute.manager [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 642.276639] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f44d9a-d18f-4406-8948-b605a5648de7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.393591] env[62368]: ERROR nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. [ 642.393591] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 642.393591] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.393591] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 642.393591] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.393591] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 642.393591] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.393591] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 642.393591] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.393591] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 642.393591] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.393591] env[62368]: ERROR nova.compute.manager raise self.value [ 642.393591] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.393591] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 642.393591] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.393591] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 642.394168] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.394168] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 642.394168] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. [ 642.394168] env[62368]: ERROR nova.compute.manager [ 642.394168] env[62368]: Traceback (most recent call last): [ 642.394168] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 642.394168] env[62368]: listener.cb(fileno) [ 642.394168] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.394168] env[62368]: result = function(*args, **kwargs) [ 642.394168] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.394168] env[62368]: return func(*args, **kwargs) [ 642.394168] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.394168] env[62368]: raise e [ 642.394168] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.394168] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 642.394168] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.394168] env[62368]: created_port_ids = self._update_ports_for_instance( [ 642.394168] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.394168] env[62368]: with excutils.save_and_reraise_exception(): [ 642.394168] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.394168] env[62368]: self.force_reraise() [ 642.394168] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.394168] env[62368]: raise self.value [ 642.394168] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.394168] env[62368]: updated_port = self._update_port( [ 642.394168] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.394168] env[62368]: _ensure_no_port_binding_failure(port) [ 642.394168] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.394168] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 642.395529] env[62368]: nova.exception.PortBindingFailed: Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. [ 642.395529] env[62368]: Removing descriptor: 15 [ 642.459191] env[62368]: DEBUG nova.network.neutron [-] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.491110] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f61d3b36-276e-4d0f-90ef-844ac89955d9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.498669] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a253a674-24ff-47dc-a6a7-96115cbd1170 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.531159] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 642.535689] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcafd142-9e65-437b-ad4e-b1e45c5ee213 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.544126] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b569c1-4fcc-419d-abd3-640a44c84a30 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.561022] env[62368]: DEBUG nova.compute.provider_tree [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 642.564828] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.564828] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.564828] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.565045] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.565045] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.565045] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.565134] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.565251] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.565414] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.565666] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.565732] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.566476] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eda3ee0-7595-473b-b2e7-d7c25b209fad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.574987] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036f01c9-9d0f-47af-b4f2-1e13148a5072 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.588733] env[62368]: ERROR nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Traceback (most recent call last): [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] yield resources [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self.driver.spawn(context, instance, image_meta, [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] vm_ref = self.build_virtual_machine(instance, [ 642.588733] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] for vif in network_info: [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] return self._sync_wrapper(fn, *args, **kwargs) [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self.wait() [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self[:] = self._gt.wait() [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] return self._exit_event.wait() [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.589024] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] current.throw(*self._exc) [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] result = function(*args, **kwargs) [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] return func(*args, **kwargs) [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] raise e [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] nwinfo = self.network_api.allocate_for_instance( [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] created_port_ids = self._update_ports_for_instance( [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] with excutils.save_and_reraise_exception(): [ 642.589397] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self.force_reraise() [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] raise self.value [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] updated_port = self._update_port( [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] _ensure_no_port_binding_failure(port) [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] raise exception.PortBindingFailed(port_id=port['id']) [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] nova.exception.PortBindingFailed: Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. [ 642.589702] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] [ 642.589702] env[62368]: INFO nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Terminating instance [ 642.591139] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.591310] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.591478] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.794496] env[62368]: INFO nova.compute.manager [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Took 28.36 seconds to build instance. [ 642.962639] env[62368]: INFO nova.compute.manager [-] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Took 1.02 seconds to deallocate network for instance. [ 642.964936] env[62368]: DEBUG nova.compute.claims [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.965133] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.102026] env[62368]: DEBUG nova.scheduler.client.report [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 59 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 643.102348] env[62368]: DEBUG nova.compute.provider_tree [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 59 to 60 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 643.102416] env[62368]: DEBUG nova.compute.provider_tree [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 643.110325] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.117016] env[62368]: INFO nova.compute.manager [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Rebuilding instance [ 643.160249] env[62368]: DEBUG nova.compute.manager [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 643.161147] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225754c6-c188-4179-8d9a-c9bea4cd94a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.216398] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.295886] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d7d4164-505a-4eb3-badf-f5fe9dfe6f24 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "4fde6065-41a7-4a95-96df-e442762dad24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.918s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.607978] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.105s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.608662] env[62368]: ERROR nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Traceback (most recent call last): [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self.driver.spawn(context, instance, image_meta, [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] vm_ref = self.build_virtual_machine(instance, [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.608662] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] for vif in network_info: [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] return self._sync_wrapper(fn, *args, **kwargs) [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self.wait() [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self[:] = self._gt.wait() [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] return self._exit_event.wait() [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] result = hub.switch() [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 643.608950] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] return self.greenlet.switch() [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] result = function(*args, **kwargs) [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] return func(*args, **kwargs) [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] raise e [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] nwinfo = self.network_api.allocate_for_instance( [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] created_port_ids = self._update_ports_for_instance( [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] with excutils.save_and_reraise_exception(): [ 643.609276] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] self.force_reraise() [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] raise self.value [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] updated_port = self._update_port( [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] _ensure_no_port_binding_failure(port) [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] raise exception.PortBindingFailed(port_id=port['id']) [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] nova.exception.PortBindingFailed: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. [ 643.609714] env[62368]: ERROR nova.compute.manager [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] [ 643.609984] env[62368]: DEBUG nova.compute.utils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 643.611086] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.072s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.612726] env[62368]: INFO nova.compute.claims [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.615672] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Build of instance 8c179402-a881-4b89-89d7-5e9031ba78cf was re-scheduled: Binding failed for port 1c9033c1-2010-4274-82e7-7d886156b319, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 643.616123] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 643.618155] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.618351] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.618525] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 643.674447] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 643.674640] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-673ff6e8-9b16-4ef0-9565-35c0a1183304 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.682103] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 643.682103] env[62368]: value = "task-1198180" [ 643.682103] env[62368]: _type = "Task" [ 643.682103] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.690646] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198180, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.718905] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.719349] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.719566] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 643.719876] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9eb53f23-28c3-4c7a-a432-9ef6d40f0826 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.728022] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f4c4bf-c1cb-4656-8aa9-b9c3ed71f382 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.752126] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9cdf70bc-0b9b-44fa-80e7-3584fc104959 could not be found. [ 643.752379] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 643.752578] env[62368]: INFO nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Took 0.03 seconds to destroy the instance on the hypervisor. [ 643.753149] env[62368]: DEBUG oslo.service.loopingcall [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.753149] env[62368]: DEBUG nova.compute.manager [-] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.753316] env[62368]: DEBUG nova.network.neutron [-] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.781969] env[62368]: DEBUG nova.network.neutron [-] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.792178] env[62368]: DEBUG nova.compute.manager [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Received event network-changed-c6adff2e-9829-4e58-85bd-ebadf3be3437 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.792489] env[62368]: DEBUG nova.compute.manager [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Refreshing instance network info cache due to event network-changed-c6adff2e-9829-4e58-85bd-ebadf3be3437. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 643.792914] env[62368]: DEBUG oslo_concurrency.lockutils [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] Acquiring lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.796687] env[62368]: DEBUG oslo_concurrency.lockutils [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] Acquired lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.796912] env[62368]: DEBUG nova.network.neutron [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Refreshing network info cache for port c6adff2e-9829-4e58-85bd-ebadf3be3437 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 643.798692] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 644.138154] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.195150] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198180, 'name': PowerOffVM_Task, 'duration_secs': 0.105687} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.195533] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 644.195838] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 644.196936] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddd08e1-a951-4545-923f-a3bf00679732 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.206027] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 644.206273] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66097320-c14c-4d05-8bf0-4f361a8d2850 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.227284] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.232879] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 644.233233] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 644.233522] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Deleting the datastore file [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 644.233892] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17291727-ed08-450e-beb5-ed368cd0da05 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.245020] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 644.245020] env[62368]: value = "task-1198182" [ 644.245020] env[62368]: _type = "Task" [ 644.245020] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.252719] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198182, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.284941] env[62368]: DEBUG nova.network.neutron [-] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.319985] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.324132] env[62368]: DEBUG nova.network.neutron [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.403074] env[62368]: DEBUG nova.network.neutron [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.730168] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "refresh_cache-8c179402-a881-4b89-89d7-5e9031ba78cf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.730423] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 644.730603] env[62368]: DEBUG nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.731047] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 644.748849] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.754987] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198182, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.105086} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.755245] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 644.755427] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 644.755600] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 644.787289] env[62368]: INFO nova.compute.manager [-] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Took 1.03 seconds to deallocate network for instance. [ 644.789509] env[62368]: DEBUG nova.compute.claims [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.789650] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.905088] env[62368]: DEBUG oslo_concurrency.lockutils [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] Releasing lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.906377] env[62368]: DEBUG nova.compute.manager [req-516f12b5-a40d-4b43-8e23-16cf0ee7e537 req-0791df20-bcf1-415b-90f0-fd8fcba4c663 service nova] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Received event network-vif-deleted-c6adff2e-9829-4e58-85bd-ebadf3be3437 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.983261] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31781d33-4dfe-4819-8550-95d7de50a7eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.990541] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c704133-00cb-4efd-a779-c9ab47328685 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.020739] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d06548-e1ad-4cfe-a5a2-d6e565027415 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.027633] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd45f990-156c-4af2-86dc-3e4b8d81154c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.040566] env[62368]: DEBUG nova.compute.provider_tree [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.252666] env[62368]: DEBUG nova.network.neutron [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.544036] env[62368]: DEBUG nova.scheduler.client.report [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.755099] env[62368]: INFO nova.compute.manager [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 8c179402-a881-4b89-89d7-5e9031ba78cf] Took 1.02 seconds to deallocate network for instance. [ 645.794292] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.794618] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.794784] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.794968] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.795131] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.795280] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.795489] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.795645] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.795810] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.795974] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.796157] env[62368]: DEBUG nova.virt.hardware [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.797052] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ff8395-3fd4-405e-864c-0aeb657ef774 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.806069] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d6e789-160f-4d97-ad68-f0d87b8d02b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.819921] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 645.826130] env[62368]: DEBUG oslo.service.loopingcall [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.826401] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 645.826615] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56836998-1961-4a87-a087-1a943dbef6c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.844870] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 645.844870] env[62368]: value = "task-1198183" [ 645.844870] env[62368]: _type = "Task" [ 645.844870] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.855686] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198183, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.051086] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.051655] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 646.054739] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 22.025s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.355830] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198183, 'name': CreateVM_Task, 'duration_secs': 0.244308} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.356010] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 646.356444] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.356605] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.356924] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 646.357189] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-072f1e75-5eeb-48da-81fd-d97fd4aca6ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.361792] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 646.361792] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]523808ca-145e-92dc-f9a6-4925cb9a4b7a" [ 646.361792] env[62368]: _type = "Task" [ 646.361792] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.369540] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523808ca-145e-92dc-f9a6-4925cb9a4b7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.558244] env[62368]: DEBUG nova.compute.utils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 646.563956] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 646.564150] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 646.612421] env[62368]: DEBUG nova.policy [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e17ec74ebf8d4c12accd841e4503b0df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23e257a2707d4c5cb3062081be3c0546', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.787371] env[62368]: INFO nova.scheduler.client.report [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted allocations for instance 8c179402-a881-4b89-89d7-5e9031ba78cf [ 646.873745] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523808ca-145e-92dc-f9a6-4925cb9a4b7a, 'name': SearchDatastore_Task, 'duration_secs': 0.009397} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.873745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.873745] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 646.873745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.873943] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.873943] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 646.873943] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7168fcd7-2f7d-4bc1-b106-dbb9c7a41bd0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.881394] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 646.881556] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 646.882299] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ae04427-9d45-40e6-8d25-e42cc746409c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.887083] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 646.887083] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526e33d7-eb10-e307-796b-87dc32e1f955" [ 646.887083] env[62368]: _type = "Task" [ 646.887083] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.894680] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526e33d7-eb10-e307-796b-87dc32e1f955, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.913646] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Successfully created port: b88a4459-4f81-4907-af66-9b1e71db43b6 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 647.068723] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 647.098545] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance d22aff27-739f-4b0d-a0e4-b6316c252c3a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 647.098707] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 45fe27f5-0f20-48ff-ad25-03862ac9bdd9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.098831] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance bebed5b9-d067-4cf2-a203-48763685ecb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.098947] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 63c88ac7-9700-4476-9039-feb034de659d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.099076] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b0bcd561-35cd-475f-975e-e78456a87fac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.099191] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance bcecb4c6-6675-403d-9a3d-b96156b1d56a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.099303] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 4fde6065-41a7-4a95-96df-e442762dad24 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.099416] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance c9f55aeb-273b-4dae-b871-cfb1572e6535 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.099525] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 9cdf70bc-0b9b-44fa-80e7-3584fc104959 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.099633] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 8e6332f4-2b1e-404d-aa43-a13a72edd044 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.298991] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fdaa943e-66a0-449e-88ff-f5887ee6e281 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "8c179402-a881-4b89-89d7-5e9031ba78cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.465s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.398104] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526e33d7-eb10-e307-796b-87dc32e1f955, 'name': SearchDatastore_Task, 'duration_secs': 0.008293} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.398943] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a9f755a-66dd-46fa-93c2-a780eaa5a995 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.406030] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 647.406030] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e18726-275c-f52e-a45f-a97265531da7" [ 647.406030] env[62368]: _type = "Task" [ 647.406030] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.415653] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e18726-275c-f52e-a45f-a97265531da7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.603078] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 09ac5c26-9842-440c-bc14-a84c2eb23cf5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.801890] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 647.870143] env[62368]: DEBUG nova.compute.manager [req-ff602d04-a331-4fe4-b388-eaf14afe1d01 req-dd09376a-e68f-4ae4-b23c-aeb362c37304 service nova] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Received event network-changed-b88a4459-4f81-4907-af66-9b1e71db43b6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 647.871407] env[62368]: DEBUG nova.compute.manager [req-ff602d04-a331-4fe4-b388-eaf14afe1d01 req-dd09376a-e68f-4ae4-b23c-aeb362c37304 service nova] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Refreshing instance network info cache due to event network-changed-b88a4459-4f81-4907-af66-9b1e71db43b6. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 647.871407] env[62368]: DEBUG oslo_concurrency.lockutils [req-ff602d04-a331-4fe4-b388-eaf14afe1d01 req-dd09376a-e68f-4ae4-b23c-aeb362c37304 service nova] Acquiring lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.871407] env[62368]: DEBUG oslo_concurrency.lockutils [req-ff602d04-a331-4fe4-b388-eaf14afe1d01 req-dd09376a-e68f-4ae4-b23c-aeb362c37304 service nova] Acquired lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.871407] env[62368]: DEBUG nova.network.neutron [req-ff602d04-a331-4fe4-b388-eaf14afe1d01 req-dd09376a-e68f-4ae4-b23c-aeb362c37304 service nova] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Refreshing network info cache for port b88a4459-4f81-4907-af66-9b1e71db43b6 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 647.917407] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e18726-275c-f52e-a45f-a97265531da7, 'name': SearchDatastore_Task, 'duration_secs': 0.008855} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.917667] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.917928] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24/4fde6065-41a7-4a95-96df-e442762dad24.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 647.918202] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6638ce22-89ed-468b-b511-d706b35d9390 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.924618] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 647.924618] env[62368]: value = "task-1198184" [ 647.924618] env[62368]: _type = "Task" [ 647.924618] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.932350] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.031259] env[62368]: ERROR nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. [ 648.031259] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.031259] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.031259] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.031259] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.031259] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.031259] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.031259] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.031259] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.031259] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 648.031259] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.031259] env[62368]: ERROR nova.compute.manager raise self.value [ 648.031259] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.031259] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.031259] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.031259] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.031785] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.031785] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.031785] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. [ 648.031785] env[62368]: ERROR nova.compute.manager [ 648.031785] env[62368]: Traceback (most recent call last): [ 648.031785] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.031785] env[62368]: listener.cb(fileno) [ 648.031785] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.031785] env[62368]: result = function(*args, **kwargs) [ 648.031785] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.031785] env[62368]: return func(*args, **kwargs) [ 648.031785] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.031785] env[62368]: raise e [ 648.031785] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.031785] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 648.031785] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.031785] env[62368]: created_port_ids = self._update_ports_for_instance( [ 648.031785] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.031785] env[62368]: with excutils.save_and_reraise_exception(): [ 648.031785] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.031785] env[62368]: self.force_reraise() [ 648.031785] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.031785] env[62368]: raise self.value [ 648.031785] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.031785] env[62368]: updated_port = self._update_port( [ 648.031785] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.031785] env[62368]: _ensure_no_port_binding_failure(port) [ 648.031785] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.031785] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.032671] env[62368]: nova.exception.PortBindingFailed: Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. [ 648.032671] env[62368]: Removing descriptor: 15 [ 648.080979] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 648.106405] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 9518175e-e4ac-42fc-bd90-cc7e204ef2e4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.110676] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 648.110676] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 648.110676] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.110878] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 648.110878] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.110878] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 648.111088] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 648.111255] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 648.112482] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 648.112482] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 648.112482] env[62368]: DEBUG nova.virt.hardware [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 648.112680] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bfa2cc-8544-4b0d-a1ad-2ddc89636553 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.124662] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08713787-5fb7-4e04-ab3c-3cfffdc9115b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.140595] env[62368]: ERROR nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Traceback (most recent call last): [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] yield resources [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self.driver.spawn(context, instance, image_meta, [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] vm_ref = self.build_virtual_machine(instance, [ 648.140595] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] for vif in network_info: [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] return self._sync_wrapper(fn, *args, **kwargs) [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self.wait() [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self[:] = self._gt.wait() [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] return self._exit_event.wait() [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 648.140964] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] current.throw(*self._exc) [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] result = function(*args, **kwargs) [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] return func(*args, **kwargs) [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] raise e [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] nwinfo = self.network_api.allocate_for_instance( [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] created_port_ids = self._update_ports_for_instance( [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] with excutils.save_and_reraise_exception(): [ 648.141559] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self.force_reraise() [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] raise self.value [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] updated_port = self._update_port( [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] _ensure_no_port_binding_failure(port) [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] raise exception.PortBindingFailed(port_id=port['id']) [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] nova.exception.PortBindingFailed: Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. [ 648.142024] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] [ 648.142024] env[62368]: INFO nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Terminating instance [ 648.143288] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.162980] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "417f000f-cf23-404d-877c-45990d1a7c77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.163221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "417f000f-cf23-404d-877c-45990d1a7c77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.326605] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.395235] env[62368]: DEBUG nova.network.neutron [req-ff602d04-a331-4fe4-b388-eaf14afe1d01 req-dd09376a-e68f-4ae4-b23c-aeb362c37304 service nova] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.434654] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462055} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.434886] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24/4fde6065-41a7-4a95-96df-e442762dad24.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 648.435112] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 648.435365] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b28d88ed-b13d-41d3-b5fc-11feadb45e8d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.442214] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 648.442214] env[62368]: value = "task-1198185" [ 648.442214] env[62368]: _type = "Task" [ 648.442214] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.449496] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198185, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.497754] env[62368]: DEBUG nova.network.neutron [req-ff602d04-a331-4fe4-b388-eaf14afe1d01 req-dd09376a-e68f-4ae4-b23c-aeb362c37304 service nova] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.618349] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.954204] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198185, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066764} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.954623] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 648.955771] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb8f054-c093-413b-b359-40afa359314f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.976590] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24/4fde6065-41a7-4a95-96df-e442762dad24.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 648.976835] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-412e9cc4-7cb8-48a8-baff-5d224e2404d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.001106] env[62368]: DEBUG oslo_concurrency.lockutils [req-ff602d04-a331-4fe4-b388-eaf14afe1d01 req-dd09376a-e68f-4ae4-b23c-aeb362c37304 service nova] Releasing lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.001651] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.001969] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 649.005161] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 649.005161] env[62368]: value = "task-1198186" [ 649.005161] env[62368]: _type = "Task" [ 649.005161] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.018092] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198186, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.121686] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 49a8e94b-b167-41d1-b748-1a2fa4e2b940 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.517880] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198186, 'name': ReconfigVM_Task, 'duration_secs': 0.249026} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.518188] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24/4fde6065-41a7-4a95-96df-e442762dad24.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 649.518781] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4de7a15d-aeef-4cab-ab59-5077c7a8488d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.524764] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 649.524764] env[62368]: value = "task-1198187" [ 649.524764] env[62368]: _type = "Task" [ 649.524764] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.528068] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.536809] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198187, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.602657] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.624981] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance c55d3f4a-7dd7-4c98-bb72-966ed12ba03f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.920266] env[62368]: DEBUG nova.compute.manager [req-161e04ae-f18e-487c-a618-dbf9bda1e0f2 req-dcd80cc7-65c7-4279-85a3-c51c320710fc service nova] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Received event network-vif-deleted-b88a4459-4f81-4907-af66-9b1e71db43b6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.035020] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198187, 'name': Rename_Task, 'duration_secs': 0.134861} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.035310] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 650.035553] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ebad288-7602-4e80-8c54-7eb10589074d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.041666] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 650.041666] env[62368]: value = "task-1198188" [ 650.041666] env[62368]: _type = "Task" [ 650.041666] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.048882] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.105229] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.105763] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.105966] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 650.106278] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3783facf-cfa7-4cfb-ad3c-12fe966722ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.114750] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65749469-d79e-4988-b530-6ef0b07028e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.130636] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 70055136-b08b-4654-aca7-8280158d94da has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.136058] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8e6332f4-2b1e-404d-aa43-a13a72edd044 could not be found. [ 650.136280] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 650.136460] env[62368]: INFO nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Took 0.03 seconds to destroy the instance on the hypervisor. [ 650.136705] env[62368]: DEBUG oslo.service.loopingcall [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.136925] env[62368]: DEBUG nova.compute.manager [-] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.137029] env[62368]: DEBUG nova.network.neutron [-] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 650.152746] env[62368]: DEBUG nova.network.neutron [-] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.552139] env[62368]: DEBUG oslo_vmware.api [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198188, 'name': PowerOnVM_Task, 'duration_secs': 0.410258} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.552439] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 650.552648] env[62368]: DEBUG nova.compute.manager [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 650.553400] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0325c21-5f40-4344-a7fd-d968138543e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.633321] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.655050] env[62368]: DEBUG nova.network.neutron [-] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.069385] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.136411] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.157681] env[62368]: INFO nova.compute.manager [-] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Took 1.02 seconds to deallocate network for instance. [ 651.161029] env[62368]: DEBUG nova.compute.claims [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.161029] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.639078] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a7c65d2c-49c1-4c0c-b38b-4e1a89552b67 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.874223] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "4fde6065-41a7-4a95-96df-e442762dad24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.874500] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "4fde6065-41a7-4a95-96df-e442762dad24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.874765] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "4fde6065-41a7-4a95-96df-e442762dad24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.874956] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "4fde6065-41a7-4a95-96df-e442762dad24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.875142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "4fde6065-41a7-4a95-96df-e442762dad24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.879024] env[62368]: INFO nova.compute.manager [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Terminating instance [ 651.880024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "refresh_cache-4fde6065-41a7-4a95-96df-e442762dad24" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.880224] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquired lock "refresh_cache-4fde6065-41a7-4a95-96df-e442762dad24" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.880423] env[62368]: DEBUG nova.network.neutron [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 652.142811] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 38291806-d0ae-47a5-bc6e-3d0320b2bd50 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.399029] env[62368]: DEBUG nova.network.neutron [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.453651] env[62368]: DEBUG nova.network.neutron [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.646861] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a5c638d8-f761-497d-bde2-39ea90d8d151 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.955792] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Releasing lock "refresh_cache-4fde6065-41a7-4a95-96df-e442762dad24" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.956315] env[62368]: DEBUG nova.compute.manager [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.956532] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 652.957530] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c604e3-0af0-49f1-bdb8-ed38cfa9b40a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.966079] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 652.966334] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c24c4c47-2f8b-40f0-a7ca-c0f855483d41 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.972753] env[62368]: DEBUG oslo_vmware.api [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 652.972753] env[62368]: value = "task-1198189" [ 652.972753] env[62368]: _type = "Task" [ 652.972753] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.981480] env[62368]: DEBUG oslo_vmware.api [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.149604] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.482939] env[62368]: DEBUG oslo_vmware.api [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198189, 'name': PowerOffVM_Task, 'duration_secs': 0.186287} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.483258] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 653.483426] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 653.483671] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c1ea73f-345d-4166-a83c-604608aab027 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.508809] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 653.508809] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 653.508809] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Deleting the datastore file [datastore1] 4fde6065-41a7-4a95-96df-e442762dad24 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 653.508809] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68fe6230-f659-4d9e-bf77-6b221fe38392 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.515055] env[62368]: DEBUG oslo_vmware.api [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for the task: (returnval){ [ 653.515055] env[62368]: value = "task-1198191" [ 653.515055] env[62368]: _type = "Task" [ 653.515055] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.522743] env[62368]: DEBUG oslo_vmware.api [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198191, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.652830] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 3f36000e-b93d-4dda-ac39-b8459203c227 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.025551] env[62368]: DEBUG oslo_vmware.api [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Task: {'id': task-1198191, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106033} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.025827] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 654.026018] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 654.026201] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 654.026373] env[62368]: INFO nova.compute.manager [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Took 1.07 seconds to destroy the instance on the hypervisor. [ 654.026621] env[62368]: DEBUG oslo.service.loopingcall [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 654.026810] env[62368]: DEBUG nova.compute.manager [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.026904] env[62368]: DEBUG nova.network.neutron [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 654.044717] env[62368]: DEBUG nova.network.neutron [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.156276] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.547356] env[62368]: DEBUG nova.network.neutron [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.659077] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 4dc7b6b4-1bf5-4195-bb93-14756f8f9986 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.051154] env[62368]: INFO nova.compute.manager [-] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Took 1.02 seconds to deallocate network for instance. [ 655.166715] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.561449] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.668541] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b393f654-0611-493e-a89f-445dfd3f8c13 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.171889] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 085ebe93-aa24-4626-94fe-241c4297e4db has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.676038] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 13765305-2e55-4ee8-9a6f-4ae5ee724367 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.179103] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 4fe7b1a2-f894-4131-9456-fb4df8a4532d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.180099] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 657.180099] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 657.517011] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938548c1-f646-4282-bbea-e4b7f2282e55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.525048] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e389104c-f10a-4bba-a556-07902e6fc187 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.554219] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687c9ed2-f84e-4145-a0f6-bfd4047fbc19 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.561204] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7100cecc-5c6c-4d30-ab31-3bfb27de360f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.573814] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.077691] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.586546] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 658.586546] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.531s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.586546] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.509s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.519484] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e1b048-82be-47da-874f-0fc50b95a2e1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.527478] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc89bd9e-0388-4fa3-91a8-f41a0b4227cf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.560826] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7495818-56bc-48d2-9754-900fcdd38cd7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.568265] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f3213f-c228-4b61-8de9-367364134f41 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.582028] env[62368]: DEBUG nova.compute.provider_tree [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.085740] env[62368]: DEBUG nova.scheduler.client.report [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.590397] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.005s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.591053] env[62368]: ERROR nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Traceback (most recent call last): [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self.driver.spawn(context, instance, image_meta, [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] vm_ref = self.build_virtual_machine(instance, [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.591053] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] for vif in network_info: [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] return self._sync_wrapper(fn, *args, **kwargs) [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self.wait() [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self[:] = self._gt.wait() [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] return self._exit_event.wait() [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] result = hub.switch() [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.591345] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] return self.greenlet.switch() [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] result = function(*args, **kwargs) [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] return func(*args, **kwargs) [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] raise e [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] nwinfo = self.network_api.allocate_for_instance( [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] created_port_ids = self._update_ports_for_instance( [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] with excutils.save_and_reraise_exception(): [ 660.591660] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] self.force_reraise() [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] raise self.value [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] updated_port = self._update_port( [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] _ensure_no_port_binding_failure(port) [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] raise exception.PortBindingFailed(port_id=port['id']) [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] nova.exception.PortBindingFailed: Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. [ 660.591972] env[62368]: ERROR nova.compute.manager [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] [ 660.592248] env[62368]: DEBUG nova.compute.utils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.593093] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.471s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.593345] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.595245] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.063s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.598125] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Build of instance 45fe27f5-0f20-48ff-ad25-03862ac9bdd9 was re-scheduled: Binding failed for port 140db14b-33b2-42f7-b7ac-4be9491c1293, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 660.598566] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 660.599299] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.599299] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquired lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.599299] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.620237] env[62368]: INFO nova.scheduler.client.report [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Deleted allocations for instance d22aff27-739f-4b0d-a0e4-b6316c252c3a [ 661.123017] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.129185] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8513bcf-c4b2-424d-81e1-d4ae475ee197 tempest-ServerDiagnosticsV248Test-9778895 tempest-ServerDiagnosticsV248Test-9778895-project-member] Lock "d22aff27-739f-4b0d-a0e4-b6316c252c3a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.808s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.212627] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.451469] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51465175-b879-4bd4-9768-f23e9c474445 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.458779] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1958e47f-6c3b-445b-bb69-5c1000d477c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.489234] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07150c8d-5521-4ba8-b56e-b305353ef140 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.496621] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a927e7b8-0a32-4a17-8d74-3b25f6502bbc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.509733] env[62368]: DEBUG nova.compute.provider_tree [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.720220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Releasing lock "refresh_cache-45fe27f5-0f20-48ff-ad25-03862ac9bdd9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.720220] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.720220] env[62368]: DEBUG nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.720220] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 661.745479] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.012878] env[62368]: DEBUG nova.scheduler.client.report [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.247745] env[62368]: DEBUG nova.network.neutron [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.518047] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.923s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.518690] env[62368]: ERROR nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Traceback (most recent call last): [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self.driver.spawn(context, instance, image_meta, [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] vm_ref = self.build_virtual_machine(instance, [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.518690] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] for vif in network_info: [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] return self._sync_wrapper(fn, *args, **kwargs) [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self.wait() [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self[:] = self._gt.wait() [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] return self._exit_event.wait() [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] result = hub.switch() [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.519151] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] return self.greenlet.switch() [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] result = function(*args, **kwargs) [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] return func(*args, **kwargs) [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] raise e [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] nwinfo = self.network_api.allocate_for_instance( [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] created_port_ids = self._update_ports_for_instance( [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] with excutils.save_and_reraise_exception(): [ 662.519522] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] self.force_reraise() [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] raise self.value [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] updated_port = self._update_port( [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] _ensure_no_port_binding_failure(port) [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] raise exception.PortBindingFailed(port_id=port['id']) [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] nova.exception.PortBindingFailed: Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. [ 662.519959] env[62368]: ERROR nova.compute.manager [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] [ 662.520286] env[62368]: DEBUG nova.compute.utils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 662.520755] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.052s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.524094] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Build of instance bebed5b9-d067-4cf2-a203-48763685ecb6 was re-scheduled: Binding failed for port 29c9a3fc-4255-459d-99c1-edade0cc5220, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 662.524536] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 662.524775] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.524911] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquired lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.525081] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.751050] env[62368]: INFO nova.compute.manager [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 45fe27f5-0f20-48ff-ad25-03862ac9bdd9] Took 1.03 seconds to deallocate network for instance. [ 663.062142] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.158828] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.400502] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5ce574-2f51-4bee-9f2b-fd12f26214fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.409560] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a65d11b-f3b5-40a7-b303-76875d810c60 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.443544] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dccab9-a5c6-4ca9-b996-e342e65f1123 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.451852] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13684a80-31e2-44ce-86c1-16ebbe7a0012 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.468627] env[62368]: DEBUG nova.compute.provider_tree [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.663722] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Releasing lock "refresh_cache-bebed5b9-d067-4cf2-a203-48763685ecb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.664015] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 663.665020] env[62368]: DEBUG nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.665020] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.680279] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.788817] env[62368]: INFO nova.scheduler.client.report [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Deleted allocations for instance 45fe27f5-0f20-48ff-ad25-03862ac9bdd9 [ 663.975234] env[62368]: DEBUG nova.scheduler.client.report [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.183234] env[62368]: DEBUG nova.network.neutron [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.303037] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52710e4c-38f3-4073-b37a-3b1b0a84b19d tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "45fe27f5-0f20-48ff-ad25-03862ac9bdd9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.232s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.479794] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.480473] env[62368]: ERROR nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] Traceback (most recent call last): [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self.driver.spawn(context, instance, image_meta, [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] vm_ref = self.build_virtual_machine(instance, [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.480473] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] for vif in network_info: [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] return self._sync_wrapper(fn, *args, **kwargs) [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self.wait() [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self[:] = self._gt.wait() [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] return self._exit_event.wait() [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] result = hub.switch() [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.480812] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] return self.greenlet.switch() [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] result = function(*args, **kwargs) [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] return func(*args, **kwargs) [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] raise e [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] nwinfo = self.network_api.allocate_for_instance( [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] created_port_ids = self._update_ports_for_instance( [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] with excutils.save_and_reraise_exception(): [ 664.481166] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] self.force_reraise() [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] raise self.value [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] updated_port = self._update_port( [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] _ensure_no_port_binding_failure(port) [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] raise exception.PortBindingFailed(port_id=port['id']) [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] nova.exception.PortBindingFailed: Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. [ 664.481478] env[62368]: ERROR nova.compute.manager [instance: 63c88ac7-9700-4476-9039-feb034de659d] [ 664.481751] env[62368]: DEBUG nova.compute.utils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 664.482944] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Build of instance 63c88ac7-9700-4476-9039-feb034de659d was re-scheduled: Binding failed for port d7d96866-a2f9-44a6-84e1-6fa2761c7162, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 664.483433] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 664.484032] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Acquiring lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.484032] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Acquired lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.484032] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 664.485370] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.257s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.689559] env[62368]: INFO nova.compute.manager [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: bebed5b9-d067-4cf2-a203-48763685ecb6] Took 1.02 seconds to deallocate network for instance. [ 664.811587] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.020044] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.166486] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.332140] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.412344] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6b90b2-211b-4cb0-a57e-5a63cc7643b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.420307] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3684697-82fd-4226-96cd-5d2721562569 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.456436] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b75cbc4-73b5-43bb-9385-631bc9815c01 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.468480] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63d5216-59f1-4c59-9f8c-0c4afc9066a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.485698] env[62368]: DEBUG nova.compute.provider_tree [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.602770] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.603030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.671368] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Releasing lock "refresh_cache-63c88ac7-9700-4476-9039-feb034de659d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.671543] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 665.671709] env[62368]: DEBUG nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.671872] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 665.688661] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.725271] env[62368]: INFO nova.scheduler.client.report [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Deleted allocations for instance bebed5b9-d067-4cf2-a203-48763685ecb6 [ 665.990337] env[62368]: DEBUG nova.scheduler.client.report [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.191701] env[62368]: DEBUG nova.network.neutron [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.237030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee0d660-d42b-4f24-85a6-89067acbaa2e tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "bebed5b9-d067-4cf2-a203-48763685ecb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.600s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.497088] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.012s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.497737] env[62368]: ERROR nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Traceback (most recent call last): [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self.driver.spawn(context, instance, image_meta, [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] vm_ref = self.build_virtual_machine(instance, [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.497737] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] for vif in network_info: [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] return self._sync_wrapper(fn, *args, **kwargs) [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self.wait() [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self[:] = self._gt.wait() [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] return self._exit_event.wait() [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] current.throw(*self._exc) [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.498144] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] result = function(*args, **kwargs) [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] return func(*args, **kwargs) [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] raise e [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] nwinfo = self.network_api.allocate_for_instance( [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] created_port_ids = self._update_ports_for_instance( [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] with excutils.save_and_reraise_exception(): [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] self.force_reraise() [ 666.498487] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] raise self.value [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] updated_port = self._update_port( [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] _ensure_no_port_binding_failure(port) [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] raise exception.PortBindingFailed(port_id=port['id']) [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] nova.exception.PortBindingFailed: Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. [ 666.498806] env[62368]: ERROR nova.compute.manager [instance: b0bcd561-35cd-475f-975e-e78456a87fac] [ 666.498806] env[62368]: DEBUG nova.compute.utils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 666.499705] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.289s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.502592] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Build of instance b0bcd561-35cd-475f-975e-e78456a87fac was re-scheduled: Binding failed for port 6703e135-9976-4bbf-9fad-b4cec2b0f20b, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.503019] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.503291] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquiring lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.503387] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Acquired lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.503544] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.694454] env[62368]: INFO nova.compute.manager [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] [instance: 63c88ac7-9700-4476-9039-feb034de659d] Took 1.02 seconds to deallocate network for instance. [ 666.737128] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.032974] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.106822] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.263426] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.438283] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1ce1d9-8c12-4185-881e-74b50fd7f776 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.447405] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b2bb7a-9e2d-4a4a-9f93-29976f4cf9e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.483843] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b416900-ce63-4e0f-ad9d-818d030900c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.491277] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d180a89-1fdb-4146-8067-22dd014ea01e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.504887] env[62368]: DEBUG nova.compute.provider_tree [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.613703] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Releasing lock "refresh_cache-b0bcd561-35cd-475f-975e-e78456a87fac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.616094] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.616094] env[62368]: DEBUG nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.616094] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.639505] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.728643] env[62368]: INFO nova.scheduler.client.report [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Deleted allocations for instance 63c88ac7-9700-4476-9039-feb034de659d [ 668.008016] env[62368]: DEBUG nova.scheduler.client.report [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.018187] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.018495] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.147594] env[62368]: DEBUG nova.network.neutron [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.239892] env[62368]: DEBUG oslo_concurrency.lockutils [None req-75f3ab79-460d-44e0-ba44-01e3af6fd28f tempest-FloatingIPsAssociationTestJSON-959105026 tempest-FloatingIPsAssociationTestJSON-959105026-project-member] Lock "63c88ac7-9700-4476-9039-feb034de659d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.118s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.515189] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.515189] env[62368]: ERROR nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. [ 668.515189] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Traceback (most recent call last): [ 668.515189] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.515189] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self.driver.spawn(context, instance, image_meta, [ 668.515189] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 668.515189] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.515189] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.515189] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] vm_ref = self.build_virtual_machine(instance, [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] for vif in network_info: [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] return self._sync_wrapper(fn, *args, **kwargs) [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self.wait() [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self[:] = self._gt.wait() [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] return self._exit_event.wait() [ 668.515773] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] current.throw(*self._exc) [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] result = function(*args, **kwargs) [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] return func(*args, **kwargs) [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] raise e [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] nwinfo = self.network_api.allocate_for_instance( [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] created_port_ids = self._update_ports_for_instance( [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.516231] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] with excutils.save_and_reraise_exception(): [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] self.force_reraise() [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] raise self.value [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] updated_port = self._update_port( [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] _ensure_no_port_binding_failure(port) [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] raise exception.PortBindingFailed(port_id=port['id']) [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] nova.exception.PortBindingFailed: Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. [ 668.516610] env[62368]: ERROR nova.compute.manager [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] [ 668.516956] env[62368]: DEBUG nova.compute.utils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.521209] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Build of instance bcecb4c6-6675-403d-9a3d-b96156b1d56a was re-scheduled: Binding failed for port cfc267e0-8cab-4795-8a71-ca70949ca14f, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.521209] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.521209] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Acquiring lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.521209] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Acquired lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.521376] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.521376] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.556s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.650701] env[62368]: INFO nova.compute.manager [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] [instance: b0bcd561-35cd-475f-975e-e78456a87fac] Took 1.04 seconds to deallocate network for instance. [ 668.742724] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 669.045858] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.130636] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.268711] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.456124] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248a9270-0b0f-4610-8c20-25d6eddb0b49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.464555] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf703de-7d37-418d-9221-88504111bdde {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.495680] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f637d4c5-4bdb-4d66-8053-88a1734047bc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.503183] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2566e7f-7fdd-4e11-ab39-8e3c53e778f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.516552] env[62368]: DEBUG nova.compute.provider_tree [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.636091] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Releasing lock "refresh_cache-bcecb4c6-6675-403d-9a3d-b96156b1d56a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.636383] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 669.636697] env[62368]: DEBUG nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.636756] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 669.652139] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.695681] env[62368]: INFO nova.scheduler.client.report [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Deleted allocations for instance b0bcd561-35cd-475f-975e-e78456a87fac [ 669.753732] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "300c6c32-b365-40c2-bc7d-6ce028f10476" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.753732] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.021029] env[62368]: DEBUG nova.scheduler.client.report [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.158806] env[62368]: DEBUG nova.network.neutron [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.203923] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2e79d851-ddf4-47a2-955a-9c4e1c528c85 tempest-DeleteServersAdminTestJSON-1829390744 tempest-DeleteServersAdminTestJSON-1829390744-project-member] Lock "b0bcd561-35cd-475f-975e-e78456a87fac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.070s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.526852] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.005s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.527529] env[62368]: ERROR nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Traceback (most recent call last): [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self.driver.spawn(context, instance, image_meta, [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] vm_ref = self.build_virtual_machine(instance, [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.527529] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] for vif in network_info: [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] return self._sync_wrapper(fn, *args, **kwargs) [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self.wait() [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self[:] = self._gt.wait() [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] return self._exit_event.wait() [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] current.throw(*self._exc) [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.528041] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] result = function(*args, **kwargs) [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] return func(*args, **kwargs) [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] raise e [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] nwinfo = self.network_api.allocate_for_instance( [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] created_port_ids = self._update_ports_for_instance( [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] with excutils.save_and_reraise_exception(): [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] self.force_reraise() [ 670.528575] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] raise self.value [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] updated_port = self._update_port( [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] _ensure_no_port_binding_failure(port) [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] raise exception.PortBindingFailed(port_id=port['id']) [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] nova.exception.PortBindingFailed: Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. [ 670.529160] env[62368]: ERROR nova.compute.manager [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] [ 670.529160] env[62368]: DEBUG nova.compute.utils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.530064] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.210s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.532037] env[62368]: INFO nova.compute.claims [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.535027] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Build of instance c9f55aeb-273b-4dae-b871-cfb1572e6535 was re-scheduled: Binding failed for port abe3f272-595d-431f-9af7-e237e3f8164c, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.535500] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.536285] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.536285] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.540193] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.661436] env[62368]: INFO nova.compute.manager [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] [instance: bcecb4c6-6675-403d-9a3d-b96156b1d56a] Took 1.02 seconds to deallocate network for instance. [ 670.710024] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 671.061996] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.165559] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.237359] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.668383] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-c9f55aeb-273b-4dae-b871-cfb1572e6535" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.668817] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.668817] env[62368]: DEBUG nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.669107] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 671.699395] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.703507] env[62368]: INFO nova.scheduler.client.report [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Deleted allocations for instance bcecb4c6-6675-403d-9a3d-b96156b1d56a [ 671.945961] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a7fbe6-be85-44eb-b28c-f33807a19239 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.955108] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a092bc2f-ac48-4e79-bab8-45f66be77984 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.998136] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d41ea2-8037-49a6-a08c-95e082db5e35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.008427] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a22e818-5874-4511-9d4f-28093b9e80c7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.024878] env[62368]: DEBUG nova.compute.provider_tree [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.210270] env[62368]: DEBUG nova.network.neutron [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.214617] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab1ccf11-4ed7-47c7-b095-f9ac6951b06a tempest-ImagesOneServerTestJSON-265441886 tempest-ImagesOneServerTestJSON-265441886-project-member] Lock "bcecb4c6-6675-403d-9a3d-b96156b1d56a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.402s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.528418] env[62368]: DEBUG nova.scheduler.client.report [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.712721] env[62368]: INFO nova.compute.manager [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: c9f55aeb-273b-4dae-b871-cfb1572e6535] Took 1.04 seconds to deallocate network for instance. [ 672.719712] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.035116] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.035116] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.038126] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.248s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.245852] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.545919] env[62368]: DEBUG nova.compute.utils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.547639] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.547808] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 673.670941] env[62368]: DEBUG nova.policy [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '86d43d611b904b1c88e925db1e7f26ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '514286cc83654bf2a7a01d6c5df2b195', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.748527] env[62368]: INFO nova.scheduler.client.report [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleted allocations for instance c9f55aeb-273b-4dae-b871-cfb1572e6535 [ 674.038923] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc7a055-adb7-4df2-8761-245a87b7a5c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.048356] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d467377b-d653-486e-8a7b-26fcf87d5def {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.052842] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.082431] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3a2949-c012-4998-b70b-e058ee172b34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.091629] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bfceaf6-e238-4d42-9767-8929294420b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.104650] env[62368]: DEBUG nova.compute.provider_tree [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.261157] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Successfully created port: 6ca9aadc-158c-4a17-9f09-8dfa60086cd4 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 674.268803] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8c500f4-f2e3-4041-9972-ab0deaffb226 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "c9f55aeb-273b-4dae-b871-cfb1572e6535" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.719s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.608314] env[62368]: DEBUG nova.scheduler.client.report [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.767986] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.070912] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.115161] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.076s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.115999] env[62368]: ERROR nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Traceback (most recent call last): [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self.driver.spawn(context, instance, image_meta, [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] vm_ref = self.build_virtual_machine(instance, [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.115999] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] for vif in network_info: [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] return self._sync_wrapper(fn, *args, **kwargs) [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self.wait() [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self[:] = self._gt.wait() [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] return self._exit_event.wait() [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] current.throw(*self._exc) [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.116510] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] result = function(*args, **kwargs) [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] return func(*args, **kwargs) [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] raise e [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] nwinfo = self.network_api.allocate_for_instance( [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] created_port_ids = self._update_ports_for_instance( [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] with excutils.save_and_reraise_exception(): [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] self.force_reraise() [ 675.116996] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] raise self.value [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] updated_port = self._update_port( [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] _ensure_no_port_binding_failure(port) [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] raise exception.PortBindingFailed(port_id=port['id']) [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] nova.exception.PortBindingFailed: Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. [ 675.117359] env[62368]: ERROR nova.compute.manager [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] [ 675.117359] env[62368]: DEBUG nova.compute.utils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.121228] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.121228] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.121629] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.121714] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.121890] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.122086] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.122341] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.122538] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.122747] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.123153] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.123153] env[62368]: DEBUG nova.virt.hardware [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.123511] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.797s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.125064] env[62368]: INFO nova.compute.claims [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.128500] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d41f056-00a9-4b33-b4fe-90c05083bdaa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.135034] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Build of instance 9cdf70bc-0b9b-44fa-80e7-3584fc104959 was re-scheduled: Binding failed for port c6adff2e-9829-4e58-85bd-ebadf3be3437, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.135503] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.135718] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.135864] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.136032] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.144018] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6a7709-ba15-4d1a-a875-e350852a8e82 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.252518] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "3f1af54b-392f-432a-9ffa-a133da428f94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.252518] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "3f1af54b-392f-432a-9ffa-a133da428f94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.301102] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.476209] env[62368]: DEBUG nova.compute.manager [req-04e73634-3d05-4886-b16c-5f0e26512c9f req-c79d3d69-5be0-401c-be7c-118d3ca72e21 service nova] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Received event network-changed-6ca9aadc-158c-4a17-9f09-8dfa60086cd4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.476553] env[62368]: DEBUG nova.compute.manager [req-04e73634-3d05-4886-b16c-5f0e26512c9f req-c79d3d69-5be0-401c-be7c-118d3ca72e21 service nova] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Refreshing instance network info cache due to event network-changed-6ca9aadc-158c-4a17-9f09-8dfa60086cd4. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 675.477253] env[62368]: DEBUG oslo_concurrency.lockutils [req-04e73634-3d05-4886-b16c-5f0e26512c9f req-c79d3d69-5be0-401c-be7c-118d3ca72e21 service nova] Acquiring lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.477253] env[62368]: DEBUG oslo_concurrency.lockutils [req-04e73634-3d05-4886-b16c-5f0e26512c9f req-c79d3d69-5be0-401c-be7c-118d3ca72e21 service nova] Acquired lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.477253] env[62368]: DEBUG nova.network.neutron [req-04e73634-3d05-4886-b16c-5f0e26512c9f req-c79d3d69-5be0-401c-be7c-118d3ca72e21 service nova] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Refreshing network info cache for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.639014] env[62368]: ERROR nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. [ 675.639014] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.639014] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.639014] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.639014] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.639014] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.639014] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.639014] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.639014] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.639014] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 675.639014] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.639014] env[62368]: ERROR nova.compute.manager raise self.value [ 675.639014] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.639014] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.639014] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.639014] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.639567] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.639567] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.639567] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. [ 675.639567] env[62368]: ERROR nova.compute.manager [ 675.639567] env[62368]: Traceback (most recent call last): [ 675.639567] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.639567] env[62368]: listener.cb(fileno) [ 675.639567] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.639567] env[62368]: result = function(*args, **kwargs) [ 675.639567] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.639567] env[62368]: return func(*args, **kwargs) [ 675.639567] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.639567] env[62368]: raise e [ 675.639567] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.639567] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 675.639567] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.639567] env[62368]: created_port_ids = self._update_ports_for_instance( [ 675.639567] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.639567] env[62368]: with excutils.save_and_reraise_exception(): [ 675.639567] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.639567] env[62368]: self.force_reraise() [ 675.639567] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.639567] env[62368]: raise self.value [ 675.639567] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.639567] env[62368]: updated_port = self._update_port( [ 675.639567] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.639567] env[62368]: _ensure_no_port_binding_failure(port) [ 675.639567] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.639567] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.640401] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. [ 675.640401] env[62368]: Removing descriptor: 15 [ 675.648468] env[62368]: ERROR nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Traceback (most recent call last): [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] yield resources [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self.driver.spawn(context, instance, image_meta, [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] vm_ref = self.build_virtual_machine(instance, [ 675.648468] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] for vif in network_info: [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] return self._sync_wrapper(fn, *args, **kwargs) [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self.wait() [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self[:] = self._gt.wait() [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] return self._exit_event.wait() [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.649117] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] result = hub.switch() [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] return self.greenlet.switch() [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] result = function(*args, **kwargs) [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] return func(*args, **kwargs) [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] raise e [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] nwinfo = self.network_api.allocate_for_instance( [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] created_port_ids = self._update_ports_for_instance( [ 675.649465] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] with excutils.save_and_reraise_exception(): [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self.force_reraise() [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] raise self.value [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] updated_port = self._update_port( [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] _ensure_no_port_binding_failure(port) [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] raise exception.PortBindingFailed(port_id=port['id']) [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] nova.exception.PortBindingFailed: Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. [ 675.649808] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] [ 675.650168] env[62368]: INFO nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Terminating instance [ 675.650906] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.663925] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.830237] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.010908] env[62368]: DEBUG nova.network.neutron [req-04e73634-3d05-4886-b16c-5f0e26512c9f req-c79d3d69-5be0-401c-be7c-118d3ca72e21 service nova] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.140124] env[62368]: DEBUG nova.network.neutron [req-04e73634-3d05-4886-b16c-5f0e26512c9f req-c79d3d69-5be0-401c-be7c-118d3ca72e21 service nova] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.332900] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "refresh_cache-9cdf70bc-0b9b-44fa-80e7-3584fc104959" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.333851] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.333851] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.333851] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 676.354723] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.583926] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d931326-7f4f-4c03-8faf-102537f059e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.592331] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7a1722-8872-4d53-9bee-bec52bfd7d91 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.626130] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416fac8f-02f5-4f51-b7a3-1a0cd42b0ca7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.634251] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d60027e-d76f-4e6e-ac9e-4f7377b71f86 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.647640] env[62368]: DEBUG oslo_concurrency.lockutils [req-04e73634-3d05-4886-b16c-5f0e26512c9f req-c79d3d69-5be0-401c-be7c-118d3ca72e21 service nova] Releasing lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.648442] env[62368]: DEBUG nova.compute.provider_tree [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.649670] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.653298] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.860063] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.158022] env[62368]: DEBUG nova.scheduler.client.report [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.184559] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.253336] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.365436] env[62368]: INFO nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 9cdf70bc-0b9b-44fa-80e7-3584fc104959] Took 1.03 seconds to deallocate network for instance. [ 677.528770] env[62368]: DEBUG nova.compute.manager [req-cf5f09ff-e8b0-4fa4-b004-64eae3ada841 req-80425878-19be-4e7a-b6b5-6247397783cf service nova] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Received event network-vif-deleted-6ca9aadc-158c-4a17-9f09-8dfa60086cd4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.667772] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.667772] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 677.671028] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.599s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.671028] env[62368]: DEBUG nova.objects.instance [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 677.758021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.758021] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 677.758021] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 677.758021] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff44cac8-8e85-4187-88a5-f2f24c304fc1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.767898] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85754ee-d328-4799-81b1-a0f69ae25dab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.794179] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 09ac5c26-9842-440c-bc14-a84c2eb23cf5 could not be found. [ 677.796890] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 677.796890] env[62368]: INFO nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 677.796890] env[62368]: DEBUG oslo.service.loopingcall [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.796890] env[62368]: DEBUG nova.compute.manager [-] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.796890] env[62368]: DEBUG nova.network.neutron [-] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.813100] env[62368]: DEBUG nova.network.neutron [-] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.949091] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.949705] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.177109] env[62368]: DEBUG nova.compute.utils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.179178] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.179362] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 678.244040] env[62368]: DEBUG nova.policy [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '692036ae1d684e0bb478884284e7166c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8c1e2a964a54b21a442da645ff90564', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 678.315833] env[62368]: DEBUG nova.network.neutron [-] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.403556] env[62368]: INFO nova.scheduler.client.report [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Deleted allocations for instance 9cdf70bc-0b9b-44fa-80e7-3584fc104959 [ 678.622151] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Successfully created port: 26337393-7da9-4078-a18c-d3d937dcf22a {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.680536] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5af09d-2448-443a-b9c6-7d4d2b97676d tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.681674] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.521s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.684527] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 678.819956] env[62368]: INFO nova.compute.manager [-] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Took 1.02 seconds to deallocate network for instance. [ 678.822224] env[62368]: DEBUG nova.compute.claims [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 678.822398] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.913683] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "9cdf70bc-0b9b-44fa-80e7-3584fc104959" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.858s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.116228] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "0f1bff84-1fb9-471a-b685-7d527b376b0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.116526] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "0f1bff84-1fb9-471a-b685-7d527b376b0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.419268] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.587289] env[62368]: DEBUG nova.compute.manager [req-e7dab266-c331-4f23-99a5-8806336f4b98 req-fc45ddd1-7a31-466f-a06a-93da6dace84a service nova] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Received event network-changed-26337393-7da9-4078-a18c-d3d937dcf22a {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.588358] env[62368]: DEBUG nova.compute.manager [req-e7dab266-c331-4f23-99a5-8806336f4b98 req-fc45ddd1-7a31-466f-a06a-93da6dace84a service nova] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Refreshing instance network info cache due to event network-changed-26337393-7da9-4078-a18c-d3d937dcf22a. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 679.588358] env[62368]: DEBUG oslo_concurrency.lockutils [req-e7dab266-c331-4f23-99a5-8806336f4b98 req-fc45ddd1-7a31-466f-a06a-93da6dace84a service nova] Acquiring lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.588358] env[62368]: DEBUG oslo_concurrency.lockutils [req-e7dab266-c331-4f23-99a5-8806336f4b98 req-fc45ddd1-7a31-466f-a06a-93da6dace84a service nova] Acquired lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.588358] env[62368]: DEBUG nova.network.neutron [req-e7dab266-c331-4f23-99a5-8806336f4b98 req-fc45ddd1-7a31-466f-a06a-93da6dace84a service nova] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Refreshing network info cache for port 26337393-7da9-4078-a18c-d3d937dcf22a {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 679.600559] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001deeae-5858-4c5c-a18d-c7388bc64e6d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.608817] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4128908d-9177-4f00-87e3-d5a9ebc09caa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.644121] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13eb5e2-4eaf-4614-a550-e23189a09d2a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.651347] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ec65b5-1f7e-4948-beaa-a71e3d6ae76a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.667472] env[62368]: DEBUG nova.compute.provider_tree [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.702740] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.718974] env[62368]: ERROR nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. [ 679.718974] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 679.718974] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.718974] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 679.718974] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.718974] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 679.718974] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.718974] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 679.718974] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.718974] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 679.718974] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.718974] env[62368]: ERROR nova.compute.manager raise self.value [ 679.718974] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.718974] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 679.718974] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.718974] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 679.719534] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.719534] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 679.719534] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. [ 679.719534] env[62368]: ERROR nova.compute.manager [ 679.719534] env[62368]: Traceback (most recent call last): [ 679.719534] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 679.719534] env[62368]: listener.cb(fileno) [ 679.719534] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.719534] env[62368]: result = function(*args, **kwargs) [ 679.719534] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.719534] env[62368]: return func(*args, **kwargs) [ 679.719534] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.719534] env[62368]: raise e [ 679.719534] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.719534] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 679.719534] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.719534] env[62368]: created_port_ids = self._update_ports_for_instance( [ 679.719534] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.719534] env[62368]: with excutils.save_and_reraise_exception(): [ 679.719534] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.719534] env[62368]: self.force_reraise() [ 679.719534] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.719534] env[62368]: raise self.value [ 679.719534] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.719534] env[62368]: updated_port = self._update_port( [ 679.719534] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.719534] env[62368]: _ensure_no_port_binding_failure(port) [ 679.719534] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.719534] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 679.720360] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. [ 679.720360] env[62368]: Removing descriptor: 15 [ 679.737683] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.738009] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.738181] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.738364] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.738554] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.738718] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.742021] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.742021] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.742021] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.742021] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.742021] env[62368]: DEBUG nova.virt.hardware [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.742398] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad1aa05-8621-43c4-b6dd-5e6c87c76afd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.749724] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd8fc8d-f155-498f-975a-7fc3111e85d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.763542] env[62368]: ERROR nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Traceback (most recent call last): [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] yield resources [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self.driver.spawn(context, instance, image_meta, [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] vm_ref = self.build_virtual_machine(instance, [ 679.763542] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] for vif in network_info: [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] return self._sync_wrapper(fn, *args, **kwargs) [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self.wait() [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self[:] = self._gt.wait() [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] return self._exit_event.wait() [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.763931] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] current.throw(*self._exc) [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] result = function(*args, **kwargs) [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] return func(*args, **kwargs) [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] raise e [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] nwinfo = self.network_api.allocate_for_instance( [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] created_port_ids = self._update_ports_for_instance( [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] with excutils.save_and_reraise_exception(): [ 679.764272] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self.force_reraise() [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] raise self.value [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] updated_port = self._update_port( [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] _ensure_no_port_binding_failure(port) [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] raise exception.PortBindingFailed(port_id=port['id']) [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] nova.exception.PortBindingFailed: Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. [ 679.764604] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] [ 679.764604] env[62368]: INFO nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Terminating instance [ 679.767508] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Acquiring lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.942263] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.106894] env[62368]: DEBUG nova.network.neutron [req-e7dab266-c331-4f23-99a5-8806336f4b98 req-fc45ddd1-7a31-466f-a06a-93da6dace84a service nova] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.170322] env[62368]: DEBUG nova.scheduler.client.report [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.185466] env[62368]: DEBUG nova.network.neutron [req-e7dab266-c331-4f23-99a5-8806336f4b98 req-fc45ddd1-7a31-466f-a06a-93da6dace84a service nova] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.675155] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.675723] env[62368]: ERROR nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Traceback (most recent call last): [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self.driver.spawn(context, instance, image_meta, [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] vm_ref = self.build_virtual_machine(instance, [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.675723] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] for vif in network_info: [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] return self._sync_wrapper(fn, *args, **kwargs) [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self.wait() [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self[:] = self._gt.wait() [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] return self._exit_event.wait() [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] current.throw(*self._exc) [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.676085] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] result = function(*args, **kwargs) [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] return func(*args, **kwargs) [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] raise e [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] nwinfo = self.network_api.allocate_for_instance( [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] created_port_ids = self._update_ports_for_instance( [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] with excutils.save_and_reraise_exception(): [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] self.force_reraise() [ 680.676464] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] raise self.value [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] updated_port = self._update_port( [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] _ensure_no_port_binding_failure(port) [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] raise exception.PortBindingFailed(port_id=port['id']) [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] nova.exception.PortBindingFailed: Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. [ 680.676840] env[62368]: ERROR nova.compute.manager [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] [ 680.676840] env[62368]: DEBUG nova.compute.utils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.677676] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.116s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.677897] env[62368]: DEBUG nova.objects.instance [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lazy-loading 'resources' on Instance uuid 4fde6065-41a7-4a95-96df-e442762dad24 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 680.679116] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Build of instance 8e6332f4-2b1e-404d-aa43-a13a72edd044 was re-scheduled: Binding failed for port b88a4459-4f81-4907-af66-9b1e71db43b6, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.679531] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.679747] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.679889] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.680082] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.688349] env[62368]: DEBUG oslo_concurrency.lockutils [req-e7dab266-c331-4f23-99a5-8806336f4b98 req-fc45ddd1-7a31-466f-a06a-93da6dace84a service nova] Releasing lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.688349] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Acquired lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.688349] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.199514] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.206768] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.278087] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.282041] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.497132] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d54c6d-8428-4af7-8549-93351d9054d9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.504413] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f916ea-dfe5-45a6-a173-844baddead5a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.534457] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd477502-f5a5-4354-9511-2287df771123 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.540924] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33f525a-03ff-4b06-b3e6-038a49e36432 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.553829] env[62368]: DEBUG nova.compute.provider_tree [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.610863] env[62368]: DEBUG nova.compute.manager [req-e97b19c7-45e8-4602-845b-49c3f4ed3db5 req-53f84784-8eb0-4a0b-9270-a081b061ad4c service nova] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Received event network-vif-deleted-26337393-7da9-4078-a18c-d3d937dcf22a {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.784614] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Releasing lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.785396] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 681.785396] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 681.785792] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "refresh_cache-8e6332f4-2b1e-404d-aa43-a13a72edd044" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.785990] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.786200] env[62368]: DEBUG nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.786366] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.788166] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca953ca8-06e3-4f6f-bf85-12377b493b0e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.797968] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc29f73e-1f1d-4ddb-83b4-008b13e662f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.809437] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.822423] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9518175e-e4ac-42fc-bd90-cc7e204ef2e4 could not be found. [ 681.822703] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 681.822889] env[62368]: INFO nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 681.823155] env[62368]: DEBUG oslo.service.loopingcall [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.823383] env[62368]: DEBUG nova.compute.manager [-] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.823713] env[62368]: DEBUG nova.network.neutron [-] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.839090] env[62368]: DEBUG nova.network.neutron [-] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.056947] env[62368]: DEBUG nova.scheduler.client.report [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.312502] env[62368]: DEBUG nova.network.neutron [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.342043] env[62368]: DEBUG nova.network.neutron [-] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.561458] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.563659] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.232s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.565130] env[62368]: INFO nova.compute.claims [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.581122] env[62368]: INFO nova.scheduler.client.report [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Deleted allocations for instance 4fde6065-41a7-4a95-96df-e442762dad24 [ 682.817180] env[62368]: INFO nova.compute.manager [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8e6332f4-2b1e-404d-aa43-a13a72edd044] Took 1.03 seconds to deallocate network for instance. [ 682.844577] env[62368]: INFO nova.compute.manager [-] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Took 1.02 seconds to deallocate network for instance. [ 682.851059] env[62368]: DEBUG nova.compute.claims [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 682.851762] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.089562] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b14fe2bc-c1fe-45bb-b2bc-cdf32e49dd65 tempest-ServerShowV257Test-1107165520 tempest-ServerShowV257Test-1107165520-project-member] Lock "4fde6065-41a7-4a95-96df-e442762dad24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.215s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.845588] env[62368]: INFO nova.scheduler.client.report [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Deleted allocations for instance 8e6332f4-2b1e-404d-aa43-a13a72edd044 [ 683.933808] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7e1049-14b8-4104-8dc2-25b39ded6ffb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.943118] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa147e81-f58c-4128-9054-b562b85c3311 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.976614] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f9754d-3f82-4bdc-a0fd-ff8b46bba6b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.986557] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fc9da1-a408-4cda-a66c-85da8110969a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.001222] env[62368]: DEBUG nova.compute.provider_tree [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.361668] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6e87676e-85c8-4754-b553-1bb300510c08 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "8e6332f4-2b1e-404d-aa43-a13a72edd044" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.234s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.504846] env[62368]: DEBUG nova.scheduler.client.report [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.865062] env[62368]: DEBUG nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.013106] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.013106] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.015603] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.752s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.018487] env[62368]: INFO nova.compute.claims [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.388167] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.523878] env[62368]: DEBUG nova.compute.utils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.527139] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.527310] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 685.567063] env[62368]: DEBUG nova.policy [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3285c80e9c5c4c6e9c4ed5aa94a2fe1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be7bc17de1dc403d8d7c84d5ea4035e4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.874706] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Successfully created port: 9a8c2fac-8758-4715-a684-a2c01d8ecdac {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.028553] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.104320] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.104320] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.141427] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.141639] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.446184] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96b0bfe-e245-4269-b941-79df9182acfb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.453946] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9584ee-7c2b-4b81-8f50-0c8257e92bab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.484939] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c36aa15-72d0-4bb6-a7cc-0cef9944e867 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.492168] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6e3d21-8ab8-449a-abc5-8e1f25972541 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.505472] env[62368]: DEBUG nova.compute.provider_tree [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.740079] env[62368]: DEBUG nova.compute.manager [req-e0d8f3e9-4803-40c5-8503-13dae1e3f4f9 req-4ac153a2-a493-4711-b0b4-166e0a7314fb service nova] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Received event network-changed-9a8c2fac-8758-4715-a684-a2c01d8ecdac {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.740079] env[62368]: DEBUG nova.compute.manager [req-e0d8f3e9-4803-40c5-8503-13dae1e3f4f9 req-4ac153a2-a493-4711-b0b4-166e0a7314fb service nova] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Refreshing instance network info cache due to event network-changed-9a8c2fac-8758-4715-a684-a2c01d8ecdac. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 686.740079] env[62368]: DEBUG oslo_concurrency.lockutils [req-e0d8f3e9-4803-40c5-8503-13dae1e3f4f9 req-4ac153a2-a493-4711-b0b4-166e0a7314fb service nova] Acquiring lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.740079] env[62368]: DEBUG oslo_concurrency.lockutils [req-e0d8f3e9-4803-40c5-8503-13dae1e3f4f9 req-4ac153a2-a493-4711-b0b4-166e0a7314fb service nova] Acquired lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.740079] env[62368]: DEBUG nova.network.neutron [req-e0d8f3e9-4803-40c5-8503-13dae1e3f4f9 req-4ac153a2-a493-4711-b0b4-166e0a7314fb service nova] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Refreshing network info cache for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 686.900317] env[62368]: ERROR nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. [ 686.900317] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.900317] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.900317] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 686.900317] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 686.900317] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 686.900317] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 686.900317] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 686.900317] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.900317] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 686.900317] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.900317] env[62368]: ERROR nova.compute.manager raise self.value [ 686.900317] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 686.900317] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 686.900317] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.900317] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 686.900969] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.900969] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 686.900969] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. [ 686.900969] env[62368]: ERROR nova.compute.manager [ 686.900969] env[62368]: Traceback (most recent call last): [ 686.900969] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 686.900969] env[62368]: listener.cb(fileno) [ 686.900969] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.900969] env[62368]: result = function(*args, **kwargs) [ 686.900969] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.900969] env[62368]: return func(*args, **kwargs) [ 686.900969] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.900969] env[62368]: raise e [ 686.900969] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.900969] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 686.900969] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 686.900969] env[62368]: created_port_ids = self._update_ports_for_instance( [ 686.900969] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 686.900969] env[62368]: with excutils.save_and_reraise_exception(): [ 686.900969] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.900969] env[62368]: self.force_reraise() [ 686.900969] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.900969] env[62368]: raise self.value [ 686.900969] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 686.900969] env[62368]: updated_port = self._update_port( [ 686.900969] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.900969] env[62368]: _ensure_no_port_binding_failure(port) [ 686.900969] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.900969] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 686.901806] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. [ 686.901806] env[62368]: Removing descriptor: 15 [ 687.008632] env[62368]: DEBUG nova.scheduler.client.report [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.044238] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.085428] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.085679] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.085828] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.086015] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.086271] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.086464] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.086709] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.086910] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.087128] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.087341] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.087553] env[62368]: DEBUG nova.virt.hardware [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.088458] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc66a58f-3e45-4866-ab2b-bc74931e73d9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.096984] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837c6506-152a-4eeb-a767-08da1cd8217b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.111636] env[62368]: ERROR nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Traceback (most recent call last): [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] yield resources [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self.driver.spawn(context, instance, image_meta, [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] vm_ref = self.build_virtual_machine(instance, [ 687.111636] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] for vif in network_info: [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] return self._sync_wrapper(fn, *args, **kwargs) [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self.wait() [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self[:] = self._gt.wait() [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] return self._exit_event.wait() [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.111940] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] current.throw(*self._exc) [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] result = function(*args, **kwargs) [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] return func(*args, **kwargs) [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] raise e [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] nwinfo = self.network_api.allocate_for_instance( [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] created_port_ids = self._update_ports_for_instance( [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] with excutils.save_and_reraise_exception(): [ 687.112270] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self.force_reraise() [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] raise self.value [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] updated_port = self._update_port( [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] _ensure_no_port_binding_failure(port) [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] raise exception.PortBindingFailed(port_id=port['id']) [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] nova.exception.PortBindingFailed: Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. [ 687.112632] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] [ 687.112632] env[62368]: INFO nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Terminating instance [ 687.114231] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.256289] env[62368]: DEBUG nova.network.neutron [req-e0d8f3e9-4803-40c5-8503-13dae1e3f4f9 req-4ac153a2-a493-4711-b0b4-166e0a7314fb service nova] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.351910] env[62368]: DEBUG nova.network.neutron [req-e0d8f3e9-4803-40c5-8503-13dae1e3f4f9 req-4ac153a2-a493-4711-b0b4-166e0a7314fb service nova] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.516882] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.517431] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.520048] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.251s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.521400] env[62368]: INFO nova.compute.claims [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.854889] env[62368]: DEBUG oslo_concurrency.lockutils [req-e0d8f3e9-4803-40c5-8503-13dae1e3f4f9 req-4ac153a2-a493-4711-b0b4-166e0a7314fb service nova] Releasing lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.855384] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquired lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.855573] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 688.025740] env[62368]: DEBUG nova.compute.utils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 688.029445] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 688.029612] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 688.072019] env[62368]: DEBUG nova.policy [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3285c80e9c5c4c6e9c4ed5aa94a2fe1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be7bc17de1dc403d8d7c84d5ea4035e4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.375555] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.386686] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Successfully created port: 1ecee59f-d22e-4cd3-970f-0535d4287492 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.476675] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.530200] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.860180] env[62368]: DEBUG nova.compute.manager [req-3e351cd0-1e9a-4374-8207-d5b5c005dd7c req-42b8aa3e-1654-4634-908c-fa1167c484df service nova] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Received event network-vif-deleted-9a8c2fac-8758-4715-a684-a2c01d8ecdac {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.944604] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a32ed24-f53e-436a-b0ae-ec75198119cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.952752] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e57d8f6-12c5-472c-b307-a86831adb52b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.988392] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Releasing lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.988392] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 688.988392] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 688.988392] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a2c74fd-08ef-41dc-b3e9-cfab3524d193 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.989504] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e18e90-0d65-4916-bcda-4366c5f1cb78 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.999838] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f3ae02-874e-460d-b8fa-20c3140328c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.007787] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a858cc-63b6-487f-8d7a-eacc0f9f3438 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.042735] env[62368]: DEBUG nova.compute.provider_tree [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.049038] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42 could not be found. [ 689.049038] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 689.049038] env[62368]: INFO nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Took 0.06 seconds to destroy the instance on the hypervisor. [ 689.049038] env[62368]: DEBUG oslo.service.loopingcall [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.049038] env[62368]: DEBUG nova.compute.manager [-] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.049038] env[62368]: DEBUG nova.network.neutron [-] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 689.065674] env[62368]: DEBUG nova.network.neutron [-] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.400372] env[62368]: ERROR nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. [ 689.400372] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.400372] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.400372] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.400372] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.400372] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.400372] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.400372] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.400372] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.400372] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 689.400372] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.400372] env[62368]: ERROR nova.compute.manager raise self.value [ 689.400372] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.400372] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.400372] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.400372] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.400897] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.400897] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.400897] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. [ 689.400897] env[62368]: ERROR nova.compute.manager [ 689.400897] env[62368]: Traceback (most recent call last): [ 689.400897] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.400897] env[62368]: listener.cb(fileno) [ 689.400897] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.400897] env[62368]: result = function(*args, **kwargs) [ 689.400897] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.400897] env[62368]: return func(*args, **kwargs) [ 689.400897] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.400897] env[62368]: raise e [ 689.400897] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.400897] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 689.400897] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.400897] env[62368]: created_port_ids = self._update_ports_for_instance( [ 689.400897] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.400897] env[62368]: with excutils.save_and_reraise_exception(): [ 689.400897] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.400897] env[62368]: self.force_reraise() [ 689.400897] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.400897] env[62368]: raise self.value [ 689.400897] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.400897] env[62368]: updated_port = self._update_port( [ 689.400897] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.400897] env[62368]: _ensure_no_port_binding_failure(port) [ 689.400897] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.400897] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.401634] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. [ 689.401634] env[62368]: Removing descriptor: 15 [ 689.544469] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.549314] env[62368]: DEBUG nova.scheduler.client.report [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.570502] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.570826] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.570942] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.571166] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.571317] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.571463] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.571661] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.571812] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.572079] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.572250] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.572419] env[62368]: DEBUG nova.virt.hardware [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.573024] env[62368]: DEBUG nova.network.neutron [-] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.575368] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7ef44e-e0cb-4b5a-8ae8-d849cfe430c7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.583513] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6452b94d-a06e-44a0-bd70-4eb2f99424f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.597557] env[62368]: ERROR nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Traceback (most recent call last): [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] yield resources [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self.driver.spawn(context, instance, image_meta, [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] vm_ref = self.build_virtual_machine(instance, [ 689.597557] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] for vif in network_info: [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] return self._sync_wrapper(fn, *args, **kwargs) [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self.wait() [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self[:] = self._gt.wait() [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] return self._exit_event.wait() [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 689.597997] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] current.throw(*self._exc) [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] result = function(*args, **kwargs) [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] return func(*args, **kwargs) [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] raise e [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] nwinfo = self.network_api.allocate_for_instance( [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] created_port_ids = self._update_ports_for_instance( [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] with excutils.save_and_reraise_exception(): [ 689.598400] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self.force_reraise() [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] raise self.value [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] updated_port = self._update_port( [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] _ensure_no_port_binding_failure(port) [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] raise exception.PortBindingFailed(port_id=port['id']) [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] nova.exception.PortBindingFailed: Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. [ 689.598779] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] [ 689.598779] env[62368]: INFO nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Terminating instance [ 689.599931] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.600106] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquired lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.600270] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.055049] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.055596] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.058211] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.821s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.059819] env[62368]: INFO nova.compute.claims [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.077917] env[62368]: INFO nova.compute.manager [-] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Took 1.03 seconds to deallocate network for instance. [ 690.083626] env[62368]: DEBUG nova.compute.claims [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 690.083778] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.117445] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.197903] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.565027] env[62368]: DEBUG nova.compute.utils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.565884] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.566068] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 690.618542] env[62368]: DEBUG nova.policy [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c632e81de6e245e9af0cbb2183586c72', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3e9e08098b54960858905c33c168884', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.700636] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Releasing lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.700919] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.701156] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 690.701462] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-875a3680-5229-42dd-8bc8-92e2be1a51ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.710324] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a8b360-ff71-48e8-92ff-d7f87f256b3e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.731393] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 49a8e94b-b167-41d1-b748-1a2fa4e2b940 could not be found. [ 690.731600] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 690.732165] env[62368]: INFO nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Took 0.03 seconds to destroy the instance on the hypervisor. [ 690.732165] env[62368]: DEBUG oslo.service.loopingcall [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.732309] env[62368]: DEBUG nova.compute.manager [-] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.732309] env[62368]: DEBUG nova.network.neutron [-] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 690.746401] env[62368]: DEBUG nova.network.neutron [-] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.915667] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Successfully created port: 2e90816a-cc3d-4e47-acbe-974abcd95bf9 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.924356] env[62368]: DEBUG nova.compute.manager [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Received event network-changed-1ecee59f-d22e-4cd3-970f-0535d4287492 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 690.924543] env[62368]: DEBUG nova.compute.manager [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Refreshing instance network info cache due to event network-changed-1ecee59f-d22e-4cd3-970f-0535d4287492. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 690.924752] env[62368]: DEBUG oslo_concurrency.lockutils [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] Acquiring lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.924894] env[62368]: DEBUG oslo_concurrency.lockutils [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] Acquired lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.925066] env[62368]: DEBUG nova.network.neutron [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Refreshing network info cache for port 1ecee59f-d22e-4cd3-970f-0535d4287492 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 691.070607] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.249207] env[62368]: DEBUG nova.network.neutron [-] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.442721] env[62368]: DEBUG nova.network.neutron [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.481853] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4db3a4-a5a7-45ab-bd2a-7487d3148c9c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.491166] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af3a42b-60e6-4fe4-98a9-a1ecac880762 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.522796] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d72cc570-5a71-48a0-8f3b-3f9ea608ff0e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.530484] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c2d1f0-24b9-4213-9025-494022bd134b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.543508] env[62368]: DEBUG nova.compute.provider_tree [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.562979] env[62368]: DEBUG nova.network.neutron [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.756069] env[62368]: INFO nova.compute.manager [-] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Took 1.02 seconds to deallocate network for instance. [ 691.760282] env[62368]: DEBUG nova.compute.claims [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.760544] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.932230] env[62368]: ERROR nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. [ 691.932230] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.932230] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.932230] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.932230] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.932230] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.932230] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.932230] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.932230] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.932230] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 691.932230] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.932230] env[62368]: ERROR nova.compute.manager raise self.value [ 691.932230] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.932230] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.932230] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.932230] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.932648] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.932648] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.932648] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. [ 691.932648] env[62368]: ERROR nova.compute.manager [ 691.932648] env[62368]: Traceback (most recent call last): [ 691.932648] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.932648] env[62368]: listener.cb(fileno) [ 691.932648] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.932648] env[62368]: result = function(*args, **kwargs) [ 691.932648] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.932648] env[62368]: return func(*args, **kwargs) [ 691.932648] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.932648] env[62368]: raise e [ 691.932648] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.932648] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 691.932648] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.932648] env[62368]: created_port_ids = self._update_ports_for_instance( [ 691.932648] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.932648] env[62368]: with excutils.save_and_reraise_exception(): [ 691.932648] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.932648] env[62368]: self.force_reraise() [ 691.932648] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.932648] env[62368]: raise self.value [ 691.932648] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.932648] env[62368]: updated_port = self._update_port( [ 691.932648] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.932648] env[62368]: _ensure_no_port_binding_failure(port) [ 691.932648] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.932648] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.933795] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. [ 691.933795] env[62368]: Removing descriptor: 15 [ 692.046940] env[62368]: DEBUG nova.scheduler.client.report [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.065832] env[62368]: DEBUG oslo_concurrency.lockutils [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] Releasing lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.066108] env[62368]: DEBUG nova.compute.manager [req-0169258e-f2fd-4ee7-a4d7-508512d70126 req-81996658-d9b7-43d2-a3de-c6a1e4be69fc service nova] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Received event network-vif-deleted-1ecee59f-d22e-4cd3-970f-0535d4287492 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.081708] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.107696] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.108027] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.108114] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.108293] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.108435] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.108577] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.108797] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.108926] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.109104] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.109266] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.109436] env[62368]: DEBUG nova.virt.hardware [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.110297] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d953f15-2556-42ec-8eba-89c5b64d8b34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.118149] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5609119e-440d-4525-a945-5cb56ac8e455 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.132205] env[62368]: ERROR nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Traceback (most recent call last): [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] yield resources [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self.driver.spawn(context, instance, image_meta, [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] vm_ref = self.build_virtual_machine(instance, [ 692.132205] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] for vif in network_info: [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] return self._sync_wrapper(fn, *args, **kwargs) [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self.wait() [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self[:] = self._gt.wait() [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] return self._exit_event.wait() [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 692.132632] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] current.throw(*self._exc) [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] result = function(*args, **kwargs) [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] return func(*args, **kwargs) [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] raise e [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] nwinfo = self.network_api.allocate_for_instance( [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] created_port_ids = self._update_ports_for_instance( [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] with excutils.save_and_reraise_exception(): [ 692.133052] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self.force_reraise() [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] raise self.value [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] updated_port = self._update_port( [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] _ensure_no_port_binding_failure(port) [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] raise exception.PortBindingFailed(port_id=port['id']) [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] nova.exception.PortBindingFailed: Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. [ 692.133482] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] [ 692.133482] env[62368]: INFO nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Terminating instance [ 692.134558] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Acquiring lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.134725] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Acquired lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.134894] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 692.551985] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.552676] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.556616] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.311s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.559072] env[62368]: INFO nova.compute.claims [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.651984] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.749554] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.953545] env[62368]: DEBUG nova.compute.manager [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Received event network-changed-2e90816a-cc3d-4e47-acbe-974abcd95bf9 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.953761] env[62368]: DEBUG nova.compute.manager [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Refreshing instance network info cache due to event network-changed-2e90816a-cc3d-4e47-acbe-974abcd95bf9. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 692.953952] env[62368]: DEBUG oslo_concurrency.lockutils [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] Acquiring lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.058499] env[62368]: DEBUG nova.compute.utils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.060101] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.060383] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.113909] env[62368]: DEBUG nova.policy [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aa18c776b1c04a3896e87288defcdb23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37024bc75bf644b19ce86a8ddb1be61e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.253799] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Releasing lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.254121] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 693.254310] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 693.254627] env[62368]: DEBUG oslo_concurrency.lockutils [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] Acquired lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.255019] env[62368]: DEBUG nova.network.neutron [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Refreshing network info cache for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 693.255974] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70124240-818b-41bf-aa26-8a6aaf834529 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.273481] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd7b535-70f7-40a0-9e06-176ef41c049d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.297196] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c55d3f4a-7dd7-4c98-bb72-966ed12ba03f could not be found. [ 693.297196] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 693.297196] env[62368]: INFO nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 693.297423] env[62368]: DEBUG oslo.service.loopingcall [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.297637] env[62368]: DEBUG nova.compute.manager [-] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.298243] env[62368]: DEBUG nova.network.neutron [-] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 693.314424] env[62368]: DEBUG nova.network.neutron [-] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.417591] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Successfully created port: 1f06cfe6-31e5-409e-be4f-d8f5463e8177 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.564081] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.778906] env[62368]: DEBUG nova.network.neutron [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.816706] env[62368]: DEBUG nova.network.neutron [-] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.868242] env[62368]: DEBUG nova.network.neutron [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.953255] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9caab1cd-f1bf-419d-afbf-23df38b1d14d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.961285] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90dd4ca-2faf-4a46-b83b-1ffbc3e39141 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.994236] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308d16b9-152e-463a-9801-fe38908c9977 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.002065] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd66b3f4-68d0-42e3-baf0-cfbaf78fcd3b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.023655] env[62368]: DEBUG nova.compute.provider_tree [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.321455] env[62368]: INFO nova.compute.manager [-] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Took 1.02 seconds to deallocate network for instance. [ 694.323993] env[62368]: DEBUG nova.compute.claims [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 694.324201] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.371387] env[62368]: DEBUG oslo_concurrency.lockutils [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] Releasing lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.371626] env[62368]: DEBUG nova.compute.manager [req-c228b009-281a-4df3-ab83-7b630b09a59a req-7a15275d-70ff-4d06-845c-e7c9ab363a10 service nova] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Received event network-vif-deleted-2e90816a-cc3d-4e47-acbe-974abcd95bf9 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.420058] env[62368]: ERROR nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. [ 694.420058] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.420058] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.420058] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.420058] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.420058] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.420058] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.420058] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.420058] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.420058] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 694.420058] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.420058] env[62368]: ERROR nova.compute.manager raise self.value [ 694.420058] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.420058] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.420058] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.420058] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.420482] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.420482] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.420482] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. [ 694.420482] env[62368]: ERROR nova.compute.manager [ 694.420482] env[62368]: Traceback (most recent call last): [ 694.420482] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.420482] env[62368]: listener.cb(fileno) [ 694.420482] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.420482] env[62368]: result = function(*args, **kwargs) [ 694.420482] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.420482] env[62368]: return func(*args, **kwargs) [ 694.420482] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.420482] env[62368]: raise e [ 694.420482] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.420482] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 694.420482] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.420482] env[62368]: created_port_ids = self._update_ports_for_instance( [ 694.420482] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.420482] env[62368]: with excutils.save_and_reraise_exception(): [ 694.420482] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.420482] env[62368]: self.force_reraise() [ 694.420482] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.420482] env[62368]: raise self.value [ 694.420482] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.420482] env[62368]: updated_port = self._update_port( [ 694.420482] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.420482] env[62368]: _ensure_no_port_binding_failure(port) [ 694.420482] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.420482] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.421352] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. [ 694.421352] env[62368]: Removing descriptor: 15 [ 694.527078] env[62368]: DEBUG nova.scheduler.client.report [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.574174] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.600697] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.601137] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.601378] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.601580] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.601731] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.601881] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.602106] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.602269] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.602431] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.602598] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.602827] env[62368]: DEBUG nova.virt.hardware [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.603720] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe51b6c6-ddff-4949-bc6c-06030a575156 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.612147] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42eb024b-3684-4c73-9cd6-ccb27368540f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.626292] env[62368]: ERROR nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] Traceback (most recent call last): [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] yield resources [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self.driver.spawn(context, instance, image_meta, [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] vm_ref = self.build_virtual_machine(instance, [ 694.626292] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] for vif in network_info: [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] return self._sync_wrapper(fn, *args, **kwargs) [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self.wait() [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self[:] = self._gt.wait() [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] return self._exit_event.wait() [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.626604] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] current.throw(*self._exc) [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] result = function(*args, **kwargs) [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] return func(*args, **kwargs) [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] raise e [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] nwinfo = self.network_api.allocate_for_instance( [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] created_port_ids = self._update_ports_for_instance( [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] with excutils.save_and_reraise_exception(): [ 694.626920] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self.force_reraise() [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] raise self.value [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] updated_port = self._update_port( [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] _ensure_no_port_binding_failure(port) [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] raise exception.PortBindingFailed(port_id=port['id']) [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] nova.exception.PortBindingFailed: Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. [ 694.627262] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] [ 694.627262] env[62368]: INFO nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Terminating instance [ 694.628896] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Acquiring lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.629855] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Acquired lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.629855] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.977353] env[62368]: DEBUG nova.compute.manager [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] [instance: 70055136-b08b-4654-aca7-8280158d94da] Received event network-changed-1f06cfe6-31e5-409e-be4f-d8f5463e8177 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.978037] env[62368]: DEBUG nova.compute.manager [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] [instance: 70055136-b08b-4654-aca7-8280158d94da] Refreshing instance network info cache due to event network-changed-1f06cfe6-31e5-409e-be4f-d8f5463e8177. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 694.978037] env[62368]: DEBUG oslo_concurrency.lockutils [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] Acquiring lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.031590] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.032131] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.034928] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.735s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.036373] env[62368]: INFO nova.compute.claims [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.147434] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.251149] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.540728] env[62368]: DEBUG nova.compute.utils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.542253] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.542931] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 695.593368] env[62368]: DEBUG nova.policy [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1cdb27b992de4336872aed2e105e1937', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e132b3dcf864356ab11d340191e0090', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.753016] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Releasing lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.753480] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 695.753706] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 695.754047] env[62368]: DEBUG oslo_concurrency.lockutils [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] Acquired lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.754258] env[62368]: DEBUG nova.network.neutron [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] [instance: 70055136-b08b-4654-aca7-8280158d94da] Refreshing network info cache for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 695.755367] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9aa7091-9272-4530-aa07-737ffaaf2534 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.764506] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3705474e-60ad-41c0-95ab-b820164656f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.787676] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 70055136-b08b-4654-aca7-8280158d94da could not be found. [ 695.787881] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 695.788082] env[62368]: INFO nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Took 0.03 seconds to destroy the instance on the hypervisor. [ 695.788328] env[62368]: DEBUG oslo.service.loopingcall [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.788546] env[62368]: DEBUG nova.compute.manager [-] [instance: 70055136-b08b-4654-aca7-8280158d94da] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.788638] env[62368]: DEBUG nova.network.neutron [-] [instance: 70055136-b08b-4654-aca7-8280158d94da] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 695.803237] env[62368]: DEBUG nova.network.neutron [-] [instance: 70055136-b08b-4654-aca7-8280158d94da] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.047305] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 696.069997] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Successfully created port: f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.287598] env[62368]: DEBUG nova.network.neutron [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] [instance: 70055136-b08b-4654-aca7-8280158d94da] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.305250] env[62368]: DEBUG nova.network.neutron [-] [instance: 70055136-b08b-4654-aca7-8280158d94da] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.403931] env[62368]: DEBUG nova.network.neutron [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] [instance: 70055136-b08b-4654-aca7-8280158d94da] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.469303] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fc6998-3331-47c6-bc37-db200e7cb416 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.478171] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16de607c-18e2-4428-a218-05c564c875fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.508210] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490633c5-e5e5-4cae-8ac1-a88d0989fedd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.515325] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b312593-8b19-41cf-80b5-9f05e65f6636 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.530009] env[62368]: DEBUG nova.compute.provider_tree [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.808648] env[62368]: INFO nova.compute.manager [-] [instance: 70055136-b08b-4654-aca7-8280158d94da] Took 1.02 seconds to deallocate network for instance. [ 696.815747] env[62368]: DEBUG nova.compute.claims [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 696.815960] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.909918] env[62368]: DEBUG oslo_concurrency.lockutils [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] Releasing lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.910218] env[62368]: DEBUG nova.compute.manager [req-b6b9611b-f058-4735-bf9c-bc0ea4165e4f req-98f161db-1661-4130-af52-630b0f57b45c service nova] [instance: 70055136-b08b-4654-aca7-8280158d94da] Received event network-vif-deleted-1f06cfe6-31e5-409e-be4f-d8f5463e8177 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.999582] env[62368]: DEBUG nova.compute.manager [req-31d23b30-bfc8-4cf5-b822-3a8a0824f216 req-5d0e0def-f292-4f49-81d4-13525e3de412 service nova] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Received event network-changed-f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.999832] env[62368]: DEBUG nova.compute.manager [req-31d23b30-bfc8-4cf5-b822-3a8a0824f216 req-5d0e0def-f292-4f49-81d4-13525e3de412 service nova] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Refreshing instance network info cache due to event network-changed-f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 696.999992] env[62368]: DEBUG oslo_concurrency.lockutils [req-31d23b30-bfc8-4cf5-b822-3a8a0824f216 req-5d0e0def-f292-4f49-81d4-13525e3de412 service nova] Acquiring lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.000167] env[62368]: DEBUG oslo_concurrency.lockutils [req-31d23b30-bfc8-4cf5-b822-3a8a0824f216 req-5d0e0def-f292-4f49-81d4-13525e3de412 service nova] Acquired lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.000390] env[62368]: DEBUG nova.network.neutron [req-31d23b30-bfc8-4cf5-b822-3a8a0824f216 req-5d0e0def-f292-4f49-81d4-13525e3de412 service nova] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Refreshing network info cache for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 697.034024] env[62368]: DEBUG nova.scheduler.client.report [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.059817] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 697.078296] env[62368]: ERROR nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. [ 697.078296] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 697.078296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.078296] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 697.078296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.078296] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 697.078296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.078296] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 697.078296] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.078296] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 697.078296] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.078296] env[62368]: ERROR nova.compute.manager raise self.value [ 697.078296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.078296] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 697.078296] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.078296] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 697.078752] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.078752] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 697.078752] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. [ 697.078752] env[62368]: ERROR nova.compute.manager [ 697.078752] env[62368]: Traceback (most recent call last): [ 697.078752] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 697.078752] env[62368]: listener.cb(fileno) [ 697.078752] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.078752] env[62368]: result = function(*args, **kwargs) [ 697.078752] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.078752] env[62368]: return func(*args, **kwargs) [ 697.078752] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.078752] env[62368]: raise e [ 697.078752] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.078752] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 697.078752] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.078752] env[62368]: created_port_ids = self._update_ports_for_instance( [ 697.078752] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.078752] env[62368]: with excutils.save_and_reraise_exception(): [ 697.078752] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.078752] env[62368]: self.force_reraise() [ 697.078752] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.078752] env[62368]: raise self.value [ 697.078752] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.078752] env[62368]: updated_port = self._update_port( [ 697.078752] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.078752] env[62368]: _ensure_no_port_binding_failure(port) [ 697.078752] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.078752] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 697.079482] env[62368]: nova.exception.PortBindingFailed: Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. [ 697.079482] env[62368]: Removing descriptor: 15 [ 697.092285] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.092285] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.092285] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.092486] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.092561] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.092761] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.093034] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.093202] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.093428] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.093643] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.093832] env[62368]: DEBUG nova.virt.hardware [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.094895] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856c6210-0933-4ff9-b924-feea56bd6aa4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.103987] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1157e5-8e4e-4858-a7b6-5fdefa4588ef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.116191] env[62368]: ERROR nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Traceback (most recent call last): [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] yield resources [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self.driver.spawn(context, instance, image_meta, [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] vm_ref = self.build_virtual_machine(instance, [ 697.116191] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] for vif in network_info: [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] return self._sync_wrapper(fn, *args, **kwargs) [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self.wait() [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self[:] = self._gt.wait() [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] return self._exit_event.wait() [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 697.116567] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] current.throw(*self._exc) [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] result = function(*args, **kwargs) [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] return func(*args, **kwargs) [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] raise e [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] nwinfo = self.network_api.allocate_for_instance( [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] created_port_ids = self._update_ports_for_instance( [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] with excutils.save_and_reraise_exception(): [ 697.116977] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self.force_reraise() [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] raise self.value [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] updated_port = self._update_port( [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] _ensure_no_port_binding_failure(port) [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] raise exception.PortBindingFailed(port_id=port['id']) [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] nova.exception.PortBindingFailed: Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. [ 697.117366] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] [ 697.117366] env[62368]: INFO nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Terminating instance [ 697.118621] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Acquiring lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.518705] env[62368]: DEBUG nova.network.neutron [req-31d23b30-bfc8-4cf5-b822-3a8a0824f216 req-5d0e0def-f292-4f49-81d4-13525e3de412 service nova] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.540388] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.540892] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.543614] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.721s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.600949] env[62368]: DEBUG nova.network.neutron [req-31d23b30-bfc8-4cf5-b822-3a8a0824f216 req-5d0e0def-f292-4f49-81d4-13525e3de412 service nova] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.048243] env[62368]: DEBUG nova.compute.utils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 698.053325] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 698.053424] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 698.103144] env[62368]: DEBUG nova.policy [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc8d2c7d871a429596781cf9155526a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce52cc740c1148ba8e83bdf8ba281e0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 698.104698] env[62368]: DEBUG oslo_concurrency.lockutils [req-31d23b30-bfc8-4cf5-b822-3a8a0824f216 req-5d0e0def-f292-4f49-81d4-13525e3de412 service nova] Releasing lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.107011] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Acquired lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.107169] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 698.406723] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e65bba7-1cb2-4c32-9c2e-62919d7b1c81 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.414854] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b39151a-8267-428b-bcc6-cd9621454c55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.447743] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2766f438-c595-4d8c-8228-b916fd652047 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.456939] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-787b1989-fd67-470b-bee6-efb4672d086a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.468864] env[62368]: DEBUG nova.compute.provider_tree [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.480602] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Successfully created port: 29095761-1f84-4074-a398-6db40c48f31c {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.553946] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.624658] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.699851] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.971756] env[62368]: DEBUG nova.scheduler.client.report [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.030617] env[62368]: DEBUG nova.compute.manager [req-28f816d3-981b-4ded-8a10-8b1bc1867afc req-caee2b76-170b-449a-acdd-56913811b7e7 service nova] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Received event network-vif-deleted-f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 699.208993] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Releasing lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.208993] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 699.208993] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 699.208993] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aefc9854-a867-4c12-8fc9-eb6512a49a2e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.218244] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259fbf38-3301-49dc-9f51-f3f3b71db14f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.241908] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6 could not be found. [ 699.242022] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 699.242430] env[62368]: INFO nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 699.242430] env[62368]: DEBUG oslo.service.loopingcall [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.243024] env[62368]: DEBUG nova.compute.manager [-] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.243024] env[62368]: DEBUG nova.network.neutron [-] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 699.258990] env[62368]: DEBUG nova.network.neutron [-] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.471236] env[62368]: ERROR nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. [ 699.471236] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.471236] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.471236] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.471236] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.471236] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.471236] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.471236] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.471236] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.471236] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 699.471236] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.471236] env[62368]: ERROR nova.compute.manager raise self.value [ 699.471236] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.471236] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.471236] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.471236] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.471708] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.471708] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.471708] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. [ 699.471708] env[62368]: ERROR nova.compute.manager [ 699.471708] env[62368]: Traceback (most recent call last): [ 699.471708] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.471708] env[62368]: listener.cb(fileno) [ 699.471708] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.471708] env[62368]: result = function(*args, **kwargs) [ 699.471708] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.471708] env[62368]: return func(*args, **kwargs) [ 699.471708] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.471708] env[62368]: raise e [ 699.471708] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.471708] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 699.471708] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.471708] env[62368]: created_port_ids = self._update_ports_for_instance( [ 699.471708] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.471708] env[62368]: with excutils.save_and_reraise_exception(): [ 699.471708] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.471708] env[62368]: self.force_reraise() [ 699.471708] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.471708] env[62368]: raise self.value [ 699.471708] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.471708] env[62368]: updated_port = self._update_port( [ 699.471708] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.471708] env[62368]: _ensure_no_port_binding_failure(port) [ 699.471708] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.471708] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.472475] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. [ 699.472475] env[62368]: Removing descriptor: 15 [ 699.477384] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.477976] env[62368]: ERROR nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Traceback (most recent call last): [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self.driver.spawn(context, instance, image_meta, [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] vm_ref = self.build_virtual_machine(instance, [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.477976] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] for vif in network_info: [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] return self._sync_wrapper(fn, *args, **kwargs) [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self.wait() [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self[:] = self._gt.wait() [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] return self._exit_event.wait() [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] result = hub.switch() [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.478407] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] return self.greenlet.switch() [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] result = function(*args, **kwargs) [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] return func(*args, **kwargs) [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] raise e [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] nwinfo = self.network_api.allocate_for_instance( [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] created_port_ids = self._update_ports_for_instance( [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] with excutils.save_and_reraise_exception(): [ 699.478718] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] self.force_reraise() [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] raise self.value [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] updated_port = self._update_port( [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] _ensure_no_port_binding_failure(port) [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] raise exception.PortBindingFailed(port_id=port['id']) [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] nova.exception.PortBindingFailed: Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. [ 699.479306] env[62368]: ERROR nova.compute.manager [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] [ 699.479612] env[62368]: DEBUG nova.compute.utils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.479865] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.538s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.482038] env[62368]: INFO nova.compute.claims [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.483966] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Build of instance 09ac5c26-9842-440c-bc14-a84c2eb23cf5 was re-scheduled: Binding failed for port 6ca9aadc-158c-4a17-9f09-8dfa60086cd4, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 699.484417] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 699.484742] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.485062] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.485062] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.563673] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.587367] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.587608] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.587765] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.587942] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.588100] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.588246] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.588448] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.588606] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.589310] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.589310] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.589310] env[62368]: DEBUG nova.virt.hardware [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.590048] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab6d71c-726d-4107-b6da-2ce890a61b93 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.598087] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe9c607-c36f-41f0-a3ba-17dd58d7b995 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.612035] env[62368]: ERROR nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Traceback (most recent call last): [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] yield resources [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self.driver.spawn(context, instance, image_meta, [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] vm_ref = self.build_virtual_machine(instance, [ 699.612035] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] for vif in network_info: [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] return self._sync_wrapper(fn, *args, **kwargs) [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self.wait() [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self[:] = self._gt.wait() [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] return self._exit_event.wait() [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.612371] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] current.throw(*self._exc) [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] result = function(*args, **kwargs) [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] return func(*args, **kwargs) [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] raise e [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] nwinfo = self.network_api.allocate_for_instance( [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] created_port_ids = self._update_ports_for_instance( [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] with excutils.save_and_reraise_exception(): [ 699.612687] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self.force_reraise() [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] raise self.value [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] updated_port = self._update_port( [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] _ensure_no_port_binding_failure(port) [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] raise exception.PortBindingFailed(port_id=port['id']) [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] nova.exception.PortBindingFailed: Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. [ 699.613076] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] [ 699.613076] env[62368]: INFO nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Terminating instance [ 699.614416] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Acquiring lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.614572] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Acquired lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.614737] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.762083] env[62368]: DEBUG nova.network.neutron [-] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.005899] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.078147] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.132604] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.217152] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.264484] env[62368]: INFO nova.compute.manager [-] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Took 1.02 seconds to deallocate network for instance. [ 700.267245] env[62368]: DEBUG nova.compute.claims [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 700.267245] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.583885] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "refresh_cache-09ac5c26-9842-440c-bc14-a84c2eb23cf5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.584120] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 700.584284] env[62368]: DEBUG nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.585457] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 700.606873] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.719824] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Releasing lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.720268] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 700.720457] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 700.720932] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e17dacd7-1770-4032-a75f-838297ce416e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.729539] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8382ddb-8351-438b-aaca-1ae14a0deecd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.752224] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a7c65d2c-49c1-4c0c-b38b-4e1a89552b67 could not be found. [ 700.752440] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 700.752615] env[62368]: INFO nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Took 0.03 seconds to destroy the instance on the hypervisor. [ 700.752915] env[62368]: DEBUG oslo.service.loopingcall [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.755174] env[62368]: DEBUG nova.compute.manager [-] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.755277] env[62368]: DEBUG nova.network.neutron [-] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 700.772045] env[62368]: DEBUG nova.network.neutron [-] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.861640] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bccde236-6200-469d-aec2-db53182aa784 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.868897] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a63e414-12ae-4862-b86d-6a6c71ebcd88 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.899037] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8472ebc-1c15-4e7f-b90f-1a97a53ca62d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.906151] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd6c613-9215-4f21-bc68-34ac83a2abe2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.919042] env[62368]: DEBUG nova.compute.provider_tree [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.055546] env[62368]: DEBUG nova.compute.manager [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Received event network-changed-29095761-1f84-4074-a398-6db40c48f31c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.055746] env[62368]: DEBUG nova.compute.manager [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Refreshing instance network info cache due to event network-changed-29095761-1f84-4074-a398-6db40c48f31c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.055962] env[62368]: DEBUG oslo_concurrency.lockutils [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] Acquiring lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.056118] env[62368]: DEBUG oslo_concurrency.lockutils [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] Acquired lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.056278] env[62368]: DEBUG nova.network.neutron [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Refreshing network info cache for port 29095761-1f84-4074-a398-6db40c48f31c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 701.110642] env[62368]: DEBUG nova.network.neutron [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.275039] env[62368]: DEBUG nova.network.neutron [-] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.422148] env[62368]: DEBUG nova.scheduler.client.report [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.579632] env[62368]: DEBUG nova.network.neutron [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.614096] env[62368]: INFO nova.compute.manager [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: 09ac5c26-9842-440c-bc14-a84c2eb23cf5] Took 1.03 seconds to deallocate network for instance. [ 701.663043] env[62368]: DEBUG nova.network.neutron [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.777592] env[62368]: INFO nova.compute.manager [-] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Took 1.02 seconds to deallocate network for instance. [ 701.780016] env[62368]: DEBUG nova.compute.claims [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 701.780206] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.926835] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.927442] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.931748] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.080s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.165960] env[62368]: DEBUG oslo_concurrency.lockutils [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] Releasing lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.166267] env[62368]: DEBUG nova.compute.manager [req-36de2d7b-9cf7-45c1-95fa-05beff82a3b7 req-561b7f7a-768d-43c1-ba5e-38b506b5e183 service nova] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Received event network-vif-deleted-29095761-1f84-4074-a398-6db40c48f31c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.436418] env[62368]: DEBUG nova.compute.utils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.441223] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 702.441872] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 702.490078] env[62368]: DEBUG nova.policy [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38178b319463474dbab2722cc1ed9a35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c876b7b5bcb24d21b7e540fb477a7064', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.644880] env[62368]: INFO nova.scheduler.client.report [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Deleted allocations for instance 09ac5c26-9842-440c-bc14-a84c2eb23cf5 [ 702.817393] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Successfully created port: 5604a801-e04f-403a-aed5-44092a88f8ea {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.824582] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2960f6-c7be-4a94-895d-826cf9b0aee3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.831842] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6ecacb-9825-4f7e-8471-ac42a37408ac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.860356] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248ad770-dc51-431d-b371-c26732974715 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.867067] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7178d8-2e38-45b4-ba54-ab5466a25591 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.879512] env[62368]: DEBUG nova.compute.provider_tree [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.941888] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 703.157869] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7fae9d42-6396-46d6-ae68-5866b9b65f91 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "09ac5c26-9842-440c-bc14-a84c2eb23cf5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.638s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.383100] env[62368]: DEBUG nova.scheduler.client.report [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.662131] env[62368]: DEBUG nova.compute.manager [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 703.727273] env[62368]: DEBUG nova.compute.manager [req-7f8f3b41-f65c-46db-a1cf-d7616518ba20 req-ef50fab3-918e-4772-91ae-6d493efde457 service nova] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Received event network-changed-5604a801-e04f-403a-aed5-44092a88f8ea {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.727273] env[62368]: DEBUG nova.compute.manager [req-7f8f3b41-f65c-46db-a1cf-d7616518ba20 req-ef50fab3-918e-4772-91ae-6d493efde457 service nova] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Refreshing instance network info cache due to event network-changed-5604a801-e04f-403a-aed5-44092a88f8ea. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 703.728637] env[62368]: DEBUG oslo_concurrency.lockutils [req-7f8f3b41-f65c-46db-a1cf-d7616518ba20 req-ef50fab3-918e-4772-91ae-6d493efde457 service nova] Acquiring lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.728806] env[62368]: DEBUG oslo_concurrency.lockutils [req-7f8f3b41-f65c-46db-a1cf-d7616518ba20 req-ef50fab3-918e-4772-91ae-6d493efde457 service nova] Acquired lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.728974] env[62368]: DEBUG nova.network.neutron [req-7f8f3b41-f65c-46db-a1cf-d7616518ba20 req-ef50fab3-918e-4772-91ae-6d493efde457 service nova] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Refreshing network info cache for port 5604a801-e04f-403a-aed5-44092a88f8ea {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 703.760500] env[62368]: ERROR nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. [ 703.760500] env[62368]: ERROR nova.compute.manager Traceback (most recent call last): [ 703.760500] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.760500] env[62368]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 703.760500] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 703.760500] env[62368]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 703.760500] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 703.760500] env[62368]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 703.760500] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.760500] env[62368]: ERROR nova.compute.manager self.force_reraise() [ 703.760500] env[62368]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.760500] env[62368]: ERROR nova.compute.manager raise self.value [ 703.760500] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 703.760500] env[62368]: ERROR nova.compute.manager updated_port = self._update_port( [ 703.760500] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.760500] env[62368]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 703.761048] env[62368]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.761048] env[62368]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 703.761048] env[62368]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. [ 703.761048] env[62368]: ERROR nova.compute.manager [ 703.761048] env[62368]: Traceback (most recent call last): [ 703.761048] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 703.761048] env[62368]: listener.cb(fileno) [ 703.761048] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.761048] env[62368]: result = function(*args, **kwargs) [ 703.761048] env[62368]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.761048] env[62368]: return func(*args, **kwargs) [ 703.761048] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.761048] env[62368]: raise e [ 703.761048] env[62368]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.761048] env[62368]: nwinfo = self.network_api.allocate_for_instance( [ 703.761048] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 703.761048] env[62368]: created_port_ids = self._update_ports_for_instance( [ 703.761048] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 703.761048] env[62368]: with excutils.save_and_reraise_exception(): [ 703.761048] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.761048] env[62368]: self.force_reraise() [ 703.761048] env[62368]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.761048] env[62368]: raise self.value [ 703.761048] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 703.761048] env[62368]: updated_port = self._update_port( [ 703.761048] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.761048] env[62368]: _ensure_no_port_binding_failure(port) [ 703.761048] env[62368]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.761048] env[62368]: raise exception.PortBindingFailed(port_id=port['id']) [ 703.763064] env[62368]: nova.exception.PortBindingFailed: Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. [ 703.763064] env[62368]: Removing descriptor: 15 [ 703.887743] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.888384] env[62368]: ERROR nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Traceback (most recent call last): [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self.driver.spawn(context, instance, image_meta, [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] vm_ref = self.build_virtual_machine(instance, [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.888384] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] for vif in network_info: [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] return self._sync_wrapper(fn, *args, **kwargs) [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self.wait() [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self[:] = self._gt.wait() [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] return self._exit_event.wait() [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] current.throw(*self._exc) [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.888740] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] result = function(*args, **kwargs) [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] return func(*args, **kwargs) [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] raise e [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] nwinfo = self.network_api.allocate_for_instance( [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] created_port_ids = self._update_ports_for_instance( [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] with excutils.save_and_reraise_exception(): [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] self.force_reraise() [ 703.889725] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] raise self.value [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] updated_port = self._update_port( [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] _ensure_no_port_binding_failure(port) [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] raise exception.PortBindingFailed(port_id=port['id']) [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] nova.exception.PortBindingFailed: Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. [ 703.890196] env[62368]: ERROR nova.compute.manager [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] [ 703.890196] env[62368]: DEBUG nova.compute.utils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.890513] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.502s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.892355] env[62368]: INFO nova.compute.claims [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.894977] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Build of instance 9518175e-e4ac-42fc-bd90-cc7e204ef2e4 was re-scheduled: Binding failed for port 26337393-7da9-4078-a18c-d3d937dcf22a, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 703.896007] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 703.896007] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Acquiring lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.896007] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Acquired lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.896007] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 703.952651] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.975023] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.975281] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.975431] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.975608] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.975750] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.975891] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.976108] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.976263] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.976426] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.976578] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.976745] env[62368]: DEBUG nova.virt.hardware [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.977614] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f61b5c-3dd9-41d9-8350-82767c74a836 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.985343] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd776ba-f1b4-41fd-8b59-864fc19fbd2f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.999205] env[62368]: ERROR nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Traceback (most recent call last): [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] yield resources [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self.driver.spawn(context, instance, image_meta, [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] vm_ref = self.build_virtual_machine(instance, [ 703.999205] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] for vif in network_info: [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] return self._sync_wrapper(fn, *args, **kwargs) [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self.wait() [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self[:] = self._gt.wait() [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] return self._exit_event.wait() [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.999564] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] current.throw(*self._exc) [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] result = function(*args, **kwargs) [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] return func(*args, **kwargs) [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] raise e [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] nwinfo = self.network_api.allocate_for_instance( [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] created_port_ids = self._update_ports_for_instance( [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] with excutils.save_and_reraise_exception(): [ 703.999892] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self.force_reraise() [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] raise self.value [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] updated_port = self._update_port( [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] _ensure_no_port_binding_failure(port) [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] raise exception.PortBindingFailed(port_id=port['id']) [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] nova.exception.PortBindingFailed: Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. [ 704.000226] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] [ 704.000226] env[62368]: INFO nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Terminating instance [ 704.001886] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Acquiring lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.180403] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.248122] env[62368]: DEBUG nova.network.neutron [req-7f8f3b41-f65c-46db-a1cf-d7616518ba20 req-ef50fab3-918e-4772-91ae-6d493efde457 service nova] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.309997] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "a0cea538-b162-4504-ac34-803a2d5a8071" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.310292] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "a0cea538-b162-4504-ac34-803a2d5a8071" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.333131] env[62368]: DEBUG nova.network.neutron [req-7f8f3b41-f65c-46db-a1cf-d7616518ba20 req-ef50fab3-918e-4772-91ae-6d493efde457 service nova] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.418354] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.490220] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.838871] env[62368]: DEBUG oslo_concurrency.lockutils [req-7f8f3b41-f65c-46db-a1cf-d7616518ba20 req-ef50fab3-918e-4772-91ae-6d493efde457 service nova] Releasing lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.838871] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Acquired lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.838871] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 704.994630] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Releasing lock "refresh_cache-9518175e-e4ac-42fc-bd90-cc7e204ef2e4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.994630] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.994630] env[62368]: DEBUG nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.994630] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 705.013376] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.345741] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e34d57-81cb-4581-815a-d21009f46e32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.353724] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90cb87b-c2e4-4ced-ae5e-e80595425866 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.359275] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.395848] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594ede30-5a1f-4788-9035-6492fc9b7231 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.403018] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7124f6-8c66-4740-a8b1-047c921e9ddc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.416824] env[62368]: DEBUG nova.compute.provider_tree [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.467785] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.515952] env[62368]: DEBUG nova.network.neutron [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.749797] env[62368]: DEBUG nova.compute.manager [req-1356ebff-82eb-4671-8b57-c9656e0d38bb req-13cf8fbe-9bb7-4f8d-bf19-e5337d1105c9 service nova] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Received event network-vif-deleted-5604a801-e04f-403a-aed5-44092a88f8ea {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.922662] env[62368]: DEBUG nova.scheduler.client.report [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.970758] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Releasing lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.971053] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.971251] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 705.972181] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-276ca6b8-ff8f-4b32-9cb6-1e6fd7e5974c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.981155] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca40b1f8-1371-4aa8-af49-18ce006b901f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.001190] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf could not be found. [ 706.001417] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 706.001596] env[62368]: INFO nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Took 0.03 seconds to destroy the instance on the hypervisor. [ 706.001837] env[62368]: DEBUG oslo.service.loopingcall [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.002068] env[62368]: DEBUG nova.compute.manager [-] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.002165] env[62368]: DEBUG nova.network.neutron [-] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 706.015810] env[62368]: DEBUG nova.network.neutron [-] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.018803] env[62368]: INFO nova.compute.manager [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] [instance: 9518175e-e4ac-42fc-bd90-cc7e204ef2e4] Took 1.03 seconds to deallocate network for instance. [ 706.427871] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.428499] env[62368]: DEBUG nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.431706] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.348s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.517777] env[62368]: DEBUG nova.network.neutron [-] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.937336] env[62368]: DEBUG nova.compute.utils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.941977] env[62368]: DEBUG nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 706.942185] env[62368]: DEBUG nova.network.neutron [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 706.994932] env[62368]: DEBUG nova.policy [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ad21d0bf82e4010b407766a58762f4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63b4ad1ad58d4159ae57ccb452d4c1fd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.021105] env[62368]: INFO nova.compute.manager [-] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Took 1.02 seconds to deallocate network for instance. [ 707.026935] env[62368]: DEBUG nova.compute.claims [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Aborting claim: {{(pid=62368) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 707.026935] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.052471] env[62368]: INFO nova.scheduler.client.report [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Deleted allocations for instance 9518175e-e4ac-42fc-bd90-cc7e204ef2e4 [ 707.324340] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a64fa8e-127c-4a86-85c5-7fbd19608090 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.332888] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b297284-46a5-46f4-a5ae-8c7f92ec8750 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.366090] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7838d45-9d6d-4076-a884-40ebeaff7ac1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.369358] env[62368]: DEBUG nova.network.neutron [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Successfully created port: 2c8ae58a-cd9a-4b06-941a-7863e1d3855f {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.375901] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be586eb-ceb5-4855-8d8e-e177b84fd5ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.388981] env[62368]: DEBUG nova.compute.provider_tree [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.442888] env[62368]: DEBUG nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.567348] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3b0269fa-5bf9-47d3-9a1e-3e62da3c2585 tempest-ServerAddressesNegativeTestJSON-1398398785 tempest-ServerAddressesNegativeTestJSON-1398398785-project-member] Lock "9518175e-e4ac-42fc-bd90-cc7e204ef2e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.425s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.891983] env[62368]: DEBUG nova.scheduler.client.report [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.069687] env[62368]: DEBUG nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 708.398193] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.398814] env[62368]: ERROR nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Traceback (most recent call last): [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self.driver.spawn(context, instance, image_meta, [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] vm_ref = self.build_virtual_machine(instance, [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.398814] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] for vif in network_info: [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] return self._sync_wrapper(fn, *args, **kwargs) [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self.wait() [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self[:] = self._gt.wait() [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] return self._exit_event.wait() [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] current.throw(*self._exc) [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.399868] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] result = function(*args, **kwargs) [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] return func(*args, **kwargs) [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] raise e [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] nwinfo = self.network_api.allocate_for_instance( [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] created_port_ids = self._update_ports_for_instance( [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] with excutils.save_and_reraise_exception(): [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] self.force_reraise() [ 708.400509] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] raise self.value [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] updated_port = self._update_port( [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] _ensure_no_port_binding_failure(port) [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] raise exception.PortBindingFailed(port_id=port['id']) [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] nova.exception.PortBindingFailed: Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. [ 708.401145] env[62368]: ERROR nova.compute.manager [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] [ 708.401145] env[62368]: DEBUG nova.compute.utils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 708.401599] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Build of instance 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42 was re-scheduled: Binding failed for port 9a8c2fac-8758-4715-a684-a2c01d8ecdac, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 708.401858] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 708.402118] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.402307] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquired lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.402485] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 708.403973] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.643s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.452874] env[62368]: DEBUG nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.478372] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.478593] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.478741] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.478926] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.479084] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.479227] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.479434] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.479588] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.479747] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.479902] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.481469] env[62368]: DEBUG nova.virt.hardware [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.481469] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88f9768-1f09-4b93-9260-b619f10f6090 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.490739] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af0b8d6-667e-4a62-aceb-1b35d9fa06a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.594843] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.943519] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.954816] env[62368]: DEBUG nova.compute.manager [req-2cf20a4b-8eb6-4c29-8823-58d375d5ba8b req-07cbf7a0-6cfd-457f-b31c-0fe8bcfcaf64 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Received event network-vif-plugged-2c8ae58a-cd9a-4b06-941a-7863e1d3855f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.955049] env[62368]: DEBUG oslo_concurrency.lockutils [req-2cf20a4b-8eb6-4c29-8823-58d375d5ba8b req-07cbf7a0-6cfd-457f-b31c-0fe8bcfcaf64 service nova] Acquiring lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.955265] env[62368]: DEBUG oslo_concurrency.lockutils [req-2cf20a4b-8eb6-4c29-8823-58d375d5ba8b req-07cbf7a0-6cfd-457f-b31c-0fe8bcfcaf64 service nova] Lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.955430] env[62368]: DEBUG oslo_concurrency.lockutils [req-2cf20a4b-8eb6-4c29-8823-58d375d5ba8b req-07cbf7a0-6cfd-457f-b31c-0fe8bcfcaf64 service nova] Lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.955592] env[62368]: DEBUG nova.compute.manager [req-2cf20a4b-8eb6-4c29-8823-58d375d5ba8b req-07cbf7a0-6cfd-457f-b31c-0fe8bcfcaf64 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] No waiting events found dispatching network-vif-plugged-2c8ae58a-cd9a-4b06-941a-7863e1d3855f {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 708.955787] env[62368]: WARNING nova.compute.manager [req-2cf20a4b-8eb6-4c29-8823-58d375d5ba8b req-07cbf7a0-6cfd-457f-b31c-0fe8bcfcaf64 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Received unexpected event network-vif-plugged-2c8ae58a-cd9a-4b06-941a-7863e1d3855f for instance with vm_state building and task_state spawning. [ 709.101443] env[62368]: DEBUG nova.network.neutron [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Successfully updated port: 2c8ae58a-cd9a-4b06-941a-7863e1d3855f {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 709.109818] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.347673] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e397c94-55e5-4401-82b9-9b2265ecd8c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.355609] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f069595b-1f89-4525-a956-7ceeb5005cd5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.386634] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24cb6328-6bd0-4b9c-a218-07cd067bc92b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.393962] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4adeb19-fee2-47af-8db2-8a8c80dbb558 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.408974] env[62368]: DEBUG nova.compute.provider_tree [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.609888] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquiring lock "refresh_cache-38291806-d0ae-47a5-bc6e-3d0320b2bd50" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.610166] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquired lock "refresh_cache-38291806-d0ae-47a5-bc6e-3d0320b2bd50" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.610236] env[62368]: DEBUG nova.network.neutron [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.613911] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Releasing lock "refresh_cache-6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.614880] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 709.614880] env[62368]: DEBUG nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.614880] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 709.641795] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.913525] env[62368]: DEBUG nova.scheduler.client.report [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.144601] env[62368]: DEBUG nova.network.neutron [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.159461] env[62368]: DEBUG nova.network.neutron [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.406682] env[62368]: DEBUG nova.network.neutron [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Updating instance_info_cache with network_info: [{"id": "2c8ae58a-cd9a-4b06-941a-7863e1d3855f", "address": "fa:16:3e:91:99:9d", "network": {"id": "49a28198-b35f-43e5-8185-bf00613ac2c7", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-883373736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63b4ad1ad58d4159ae57ccb452d4c1fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c8ae58a-cd", "ovs_interfaceid": "2c8ae58a-cd9a-4b06-941a-7863e1d3855f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.418510] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.419145] env[62368]: ERROR nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Traceback (most recent call last): [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self.driver.spawn(context, instance, image_meta, [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] vm_ref = self.build_virtual_machine(instance, [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.419145] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] for vif in network_info: [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] return self._sync_wrapper(fn, *args, **kwargs) [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self.wait() [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self[:] = self._gt.wait() [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] return self._exit_event.wait() [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] current.throw(*self._exc) [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.419687] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] result = function(*args, **kwargs) [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] return func(*args, **kwargs) [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] raise e [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] nwinfo = self.network_api.allocate_for_instance( [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] created_port_ids = self._update_ports_for_instance( [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] with excutils.save_and_reraise_exception(): [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] self.force_reraise() [ 710.420311] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] raise self.value [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] updated_port = self._update_port( [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] _ensure_no_port_binding_failure(port) [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] raise exception.PortBindingFailed(port_id=port['id']) [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] nova.exception.PortBindingFailed: Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. [ 710.420874] env[62368]: ERROR nova.compute.manager [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] [ 710.420874] env[62368]: DEBUG nova.compute.utils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.421727] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.097s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.426063] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Build of instance 49a8e94b-b167-41d1-b748-1a2fa4e2b940 was re-scheduled: Binding failed for port 1ecee59f-d22e-4cd3-970f-0535d4287492, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.426063] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.426281] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquiring lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.426514] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Acquired lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.426662] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 710.652250] env[62368]: INFO nova.compute.manager [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42] Took 1.04 seconds to deallocate network for instance. [ 710.911479] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Releasing lock "refresh_cache-38291806-d0ae-47a5-bc6e-3d0320b2bd50" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.911808] env[62368]: DEBUG nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Instance network_info: |[{"id": "2c8ae58a-cd9a-4b06-941a-7863e1d3855f", "address": "fa:16:3e:91:99:9d", "network": {"id": "49a28198-b35f-43e5-8185-bf00613ac2c7", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-883373736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63b4ad1ad58d4159ae57ccb452d4c1fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c8ae58a-cd", "ovs_interfaceid": "2c8ae58a-cd9a-4b06-941a-7863e1d3855f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 710.912319] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:99:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd47d5e1d-e66d-4f2c-83e6-d5e78c2b767d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c8ae58a-cd9a-4b06-941a-7863e1d3855f', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 710.920450] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Creating folder: Project (63b4ad1ad58d4159ae57ccb452d4c1fd). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 710.920657] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be6333dd-a954-4f93-b479-b58f708dcada {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.932602] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Created folder: Project (63b4ad1ad58d4159ae57ccb452d4c1fd) in parent group-v259706. [ 710.932786] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Creating folder: Instances. Parent ref: group-v259723. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 710.933357] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0afc30ac-e8f2-4bd1-ae5e-6dc66e2eb134 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.941795] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Created folder: Instances in parent group-v259723. [ 710.942056] env[62368]: DEBUG oslo.service.loopingcall [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.942258] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 710.942467] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b659268f-4698-436f-a04c-9dacabd3ab2e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.960021] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.967568] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 710.967568] env[62368]: value = "task-1198194" [ 710.967568] env[62368]: _type = "Task" [ 710.967568] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.977436] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198194, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.986860] env[62368]: DEBUG nova.compute.manager [req-3cff259c-8027-4be7-9e52-ee1ecfd0675d req-2d40aac9-b635-49f9-979d-1183053cd471 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Received event network-changed-2c8ae58a-cd9a-4b06-941a-7863e1d3855f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.987074] env[62368]: DEBUG nova.compute.manager [req-3cff259c-8027-4be7-9e52-ee1ecfd0675d req-2d40aac9-b635-49f9-979d-1183053cd471 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Refreshing instance network info cache due to event network-changed-2c8ae58a-cd9a-4b06-941a-7863e1d3855f. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.987306] env[62368]: DEBUG oslo_concurrency.lockutils [req-3cff259c-8027-4be7-9e52-ee1ecfd0675d req-2d40aac9-b635-49f9-979d-1183053cd471 service nova] Acquiring lock "refresh_cache-38291806-d0ae-47a5-bc6e-3d0320b2bd50" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.987422] env[62368]: DEBUG oslo_concurrency.lockutils [req-3cff259c-8027-4be7-9e52-ee1ecfd0675d req-2d40aac9-b635-49f9-979d-1183053cd471 service nova] Acquired lock "refresh_cache-38291806-d0ae-47a5-bc6e-3d0320b2bd50" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.987623] env[62368]: DEBUG nova.network.neutron [req-3cff259c-8027-4be7-9e52-ee1ecfd0675d req-2d40aac9-b635-49f9-979d-1183053cd471 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Refreshing network info cache for port 2c8ae58a-cd9a-4b06-941a-7863e1d3855f {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 711.038473] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.272518] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05168316-f2c6-4130-add1-1e8c9f366ec9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.280927] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4348410f-a251-41e2-b016-48a6c7314705 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.312323] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f5cfd7-ee91-4025-819a-73f1f06fb4d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.319941] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e09632d-f6cd-45ea-b188-ff3202d191f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.333326] env[62368]: DEBUG nova.compute.provider_tree [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.477991] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198194, 'name': CreateVM_Task, 'duration_secs': 0.329268} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.478179] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 711.484762] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.484977] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.485384] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 711.485631] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc5dcfbb-850d-4f88-9a7a-cf133abf8939 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.491621] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 711.491621] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52619150-3115-4c59-1988-cd481fbb3c61" [ 711.491621] env[62368]: _type = "Task" [ 711.491621] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.499226] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52619150-3115-4c59-1988-cd481fbb3c61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.540830] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Releasing lock "refresh_cache-49a8e94b-b167-41d1-b748-1a2fa4e2b940" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.541118] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 711.541312] env[62368]: DEBUG nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.541484] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 711.558041] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.676252] env[62368]: INFO nova.scheduler.client.report [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Deleted allocations for instance 6a0c3d41-285d-4480-82e8-e0e9b0fb0a42 [ 711.750235] env[62368]: DEBUG nova.network.neutron [req-3cff259c-8027-4be7-9e52-ee1ecfd0675d req-2d40aac9-b635-49f9-979d-1183053cd471 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Updated VIF entry in instance network info cache for port 2c8ae58a-cd9a-4b06-941a-7863e1d3855f. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 711.750593] env[62368]: DEBUG nova.network.neutron [req-3cff259c-8027-4be7-9e52-ee1ecfd0675d req-2d40aac9-b635-49f9-979d-1183053cd471 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Updating instance_info_cache with network_info: [{"id": "2c8ae58a-cd9a-4b06-941a-7863e1d3855f", "address": "fa:16:3e:91:99:9d", "network": {"id": "49a28198-b35f-43e5-8185-bf00613ac2c7", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-883373736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63b4ad1ad58d4159ae57ccb452d4c1fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c8ae58a-cd", "ovs_interfaceid": "2c8ae58a-cd9a-4b06-941a-7863e1d3855f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.835823] env[62368]: DEBUG nova.scheduler.client.report [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.005135] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52619150-3115-4c59-1988-cd481fbb3c61, 'name': SearchDatastore_Task, 'duration_secs': 0.010922} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.005546] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.005864] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 712.006203] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.006424] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.006794] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 712.007076] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20bf25a4-95c6-44b5-b993-ea30295d09fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.016158] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 712.016525] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 712.017455] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55f40bc3-6dec-42d8-9922-8415446e676f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.024717] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 712.024717] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5245df20-77b5-45d3-ae41-d9cc13fbf70b" [ 712.024717] env[62368]: _type = "Task" [ 712.024717] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.032021] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5245df20-77b5-45d3-ae41-d9cc13fbf70b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.060817] env[62368]: DEBUG nova.network.neutron [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.183712] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e79ff412-2b3a-44a2-a138-c0fb06bb90cd tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "6a0c3d41-285d-4480-82e8-e0e9b0fb0a42" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.191s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.253438] env[62368]: DEBUG oslo_concurrency.lockutils [req-3cff259c-8027-4be7-9e52-ee1ecfd0675d req-2d40aac9-b635-49f9-979d-1183053cd471 service nova] Releasing lock "refresh_cache-38291806-d0ae-47a5-bc6e-3d0320b2bd50" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.340379] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.918s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.341031] env[62368]: ERROR nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Traceback (most recent call last): [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self.driver.spawn(context, instance, image_meta, [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] vm_ref = self.build_virtual_machine(instance, [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.341031] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] for vif in network_info: [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] return self._sync_wrapper(fn, *args, **kwargs) [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self.wait() [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self[:] = self._gt.wait() [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] return self._exit_event.wait() [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] current.throw(*self._exc) [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.341416] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] result = function(*args, **kwargs) [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] return func(*args, **kwargs) [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] raise e [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] nwinfo = self.network_api.allocate_for_instance( [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] created_port_ids = self._update_ports_for_instance( [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] with excutils.save_and_reraise_exception(): [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] self.force_reraise() [ 712.341826] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] raise self.value [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] updated_port = self._update_port( [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] _ensure_no_port_binding_failure(port) [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] raise exception.PortBindingFailed(port_id=port['id']) [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] nova.exception.PortBindingFailed: Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. [ 712.342252] env[62368]: ERROR nova.compute.manager [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] [ 712.342252] env[62368]: DEBUG nova.compute.utils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.343384] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.527s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.346957] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Build of instance c55d3f4a-7dd7-4c98-bb72-966ed12ba03f was re-scheduled: Binding failed for port 2e90816a-cc3d-4e47-acbe-974abcd95bf9, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.347399] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.347792] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Acquiring lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.347792] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Acquired lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.347959] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.534915] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5245df20-77b5-45d3-ae41-d9cc13fbf70b, 'name': SearchDatastore_Task, 'duration_secs': 0.008735} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.535694] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1514646-f6ed-43b5-a129-320e9381ffae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.540466] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 712.540466] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5210565a-f258-2b43-3642-4287cf444233" [ 712.540466] env[62368]: _type = "Task" [ 712.540466] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.547536] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5210565a-f258-2b43-3642-4287cf444233, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.563260] env[62368]: INFO nova.compute.manager [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] [instance: 49a8e94b-b167-41d1-b748-1a2fa4e2b940] Took 1.02 seconds to deallocate network for instance. [ 712.687256] env[62368]: DEBUG nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 712.976111] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.056063] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5210565a-f258-2b43-3642-4287cf444233, 'name': SearchDatastore_Task, 'duration_secs': 0.009225} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.056744] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.056881] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 38291806-d0ae-47a5-bc6e-3d0320b2bd50/38291806-d0ae-47a5-bc6e-3d0320b2bd50.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 713.057296] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd3beca0-1119-492a-9361-60e871bdd687 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.072288] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 713.072288] env[62368]: value = "task-1198195" [ 713.072288] env[62368]: _type = "Task" [ 713.072288] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.088545] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198195, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.103941] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.203078] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.205676] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1157d370-1bd4-4f97-8ed7-787992bd0f84 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.213228] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18975405-6d03-4ddc-af77-c8e555f09f40 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.246214] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d50e021-55a6-4516-b2ce-93179d1ad837 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.255838] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b6091d-d651-4957-ab23-eeda19b6caee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.272400] env[62368]: DEBUG nova.compute.provider_tree [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.585218] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198195, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486127} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.585507] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 38291806-d0ae-47a5-bc6e-3d0320b2bd50/38291806-d0ae-47a5-bc6e-3d0320b2bd50.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 713.585741] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 713.585898] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-823ff2f0-9b1f-45ec-94d9-2a5d6f6bd69b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.591957] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 713.591957] env[62368]: value = "task-1198196" [ 713.591957] env[62368]: _type = "Task" [ 713.591957] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.592746] env[62368]: INFO nova.scheduler.client.report [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Deleted allocations for instance 49a8e94b-b167-41d1-b748-1a2fa4e2b940 [ 713.606670] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Releasing lock "refresh_cache-c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.606921] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.607187] env[62368]: DEBUG nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.607448] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.609619] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198196, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.625471] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.775605] env[62368]: DEBUG nova.scheduler.client.report [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.102940] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3423cf49-b5b5-41d2-a8ba-2c2a2aae95f0 tempest-ServersAdminTestJSON-1386654138 tempest-ServersAdminTestJSON-1386654138-project-member] Lock "49a8e94b-b167-41d1-b748-1a2fa4e2b940" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.436s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.110083] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198196, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062638} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.110083] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 714.110919] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2b43ef-c058-4b9f-9d1e-40e8e15e454b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.133639] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 38291806-d0ae-47a5-bc6e-3d0320b2bd50/38291806-d0ae-47a5-bc6e-3d0320b2bd50.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 714.134064] env[62368]: DEBUG nova.network.neutron [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.135277] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcbe9413-a52b-4f82-b0fd-7253ab3b8718 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.149641] env[62368]: INFO nova.compute.manager [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] [instance: c55d3f4a-7dd7-4c98-bb72-966ed12ba03f] Took 0.54 seconds to deallocate network for instance. [ 714.158056] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 714.158056] env[62368]: value = "task-1198197" [ 714.158056] env[62368]: _type = "Task" [ 714.158056] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.165806] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198197, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.280309] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.280943] env[62368]: ERROR nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] Traceback (most recent call last): [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self.driver.spawn(context, instance, image_meta, [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] vm_ref = self.build_virtual_machine(instance, [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.280943] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] for vif in network_info: [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] return self._sync_wrapper(fn, *args, **kwargs) [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self.wait() [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self[:] = self._gt.wait() [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] return self._exit_event.wait() [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] current.throw(*self._exc) [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.281503] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] result = function(*args, **kwargs) [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] return func(*args, **kwargs) [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] raise e [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] nwinfo = self.network_api.allocate_for_instance( [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] created_port_ids = self._update_ports_for_instance( [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] with excutils.save_and_reraise_exception(): [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] self.force_reraise() [ 714.282046] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] raise self.value [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] updated_port = self._update_port( [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] _ensure_no_port_binding_failure(port) [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] raise exception.PortBindingFailed(port_id=port['id']) [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] nova.exception.PortBindingFailed: Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. [ 714.282678] env[62368]: ERROR nova.compute.manager [instance: 70055136-b08b-4654-aca7-8280158d94da] [ 714.282678] env[62368]: DEBUG nova.compute.utils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.283196] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.016s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.286071] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Build of instance 70055136-b08b-4654-aca7-8280158d94da was re-scheduled: Binding failed for port 1f06cfe6-31e5-409e-be4f-d8f5463e8177, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 714.286488] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 714.286712] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Acquiring lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.286854] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Acquired lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.287014] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.606576] env[62368]: DEBUG nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 714.667675] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198197, 'name': ReconfigVM_Task, 'duration_secs': 0.270012} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.667955] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 38291806-d0ae-47a5-bc6e-3d0320b2bd50/38291806-d0ae-47a5-bc6e-3d0320b2bd50.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 714.668607] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6bc6ccf1-eccd-4f74-a6b2-b37fc698aeee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.677020] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 714.677020] env[62368]: value = "task-1198198" [ 714.677020] env[62368]: _type = "Task" [ 714.677020] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.685447] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198198, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.814093] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.915372] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.107048] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c1277c-0e8a-40f2-9e1c-defc12e549ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.122588] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337ef63f-d1d2-4667-9dc8-92287e68d6fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.156323] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.157098] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd79e99-1e3d-4241-b03a-4ff528e75a16 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.167817] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899575b2-3b82-4db4-8456-01538d7653a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.184922] env[62368]: DEBUG nova.compute.provider_tree [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.186980] env[62368]: INFO nova.scheduler.client.report [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Deleted allocations for instance c55d3f4a-7dd7-4c98-bb72-966ed12ba03f [ 715.200108] env[62368]: DEBUG nova.scheduler.client.report [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.209681] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198198, 'name': Rename_Task, 'duration_secs': 0.128654} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.209952] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 715.210237] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad6d3de3-e6c7-498b-8da6-690cb22ad27e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.217379] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 715.217379] env[62368]: value = "task-1198199" [ 715.217379] env[62368]: _type = "Task" [ 715.217379] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.229814] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198199, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.420097] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Releasing lock "refresh_cache-70055136-b08b-4654-aca7-8280158d94da" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.420097] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 715.420097] env[62368]: DEBUG nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.420097] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 715.441499] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.706221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1695a45d-3116-4482-9641-2bd2bb8433e3 tempest-InstanceActionsNegativeTestJSON-1891540903 tempest-InstanceActionsNegativeTestJSON-1891540903-project-member] Lock "c55d3f4a-7dd7-4c98-bb72-966ed12ba03f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.545s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.706221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.422s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.706221] env[62368]: ERROR nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. [ 715.706221] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Traceback (most recent call last): [ 715.706221] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.706221] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self.driver.spawn(context, instance, image_meta, [ 715.706221] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] vm_ref = self.build_virtual_machine(instance, [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] for vif in network_info: [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] return self._sync_wrapper(fn, *args, **kwargs) [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self.wait() [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.706719] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self[:] = self._gt.wait() [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] return self._exit_event.wait() [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] current.throw(*self._exc) [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] result = function(*args, **kwargs) [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] return func(*args, **kwargs) [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] raise e [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] nwinfo = self.network_api.allocate_for_instance( [ 715.707086] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] created_port_ids = self._update_ports_for_instance( [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] with excutils.save_and_reraise_exception(): [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] self.force_reraise() [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] raise self.value [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] updated_port = self._update_port( [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] _ensure_no_port_binding_failure(port) [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.707431] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] raise exception.PortBindingFailed(port_id=port['id']) [ 715.707778] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] nova.exception.PortBindingFailed: Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. [ 715.707778] env[62368]: ERROR nova.compute.manager [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] [ 715.710359] env[62368]: DEBUG nova.compute.utils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 715.710810] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.931s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.722151] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Build of instance 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6 was re-scheduled: Binding failed for port f2a52a0d-f5bc-4b01-809b-f8d2a1b3e407, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 715.722151] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 715.722151] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Acquiring lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.722151] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Acquired lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.722151] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.731705] env[62368]: DEBUG oslo_vmware.api [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198199, 'name': PowerOnVM_Task, 'duration_secs': 0.422955} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.732768] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 715.733389] env[62368]: INFO nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Took 7.28 seconds to spawn the instance on the hypervisor. [ 715.733454] env[62368]: DEBUG nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 715.734636] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa64dce-f0da-42d4-83c7-219cd6ebb6a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.944791] env[62368]: DEBUG nova.network.neutron [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.221123] env[62368]: DEBUG nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.252360] env[62368]: INFO nova.compute.manager [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Took 30.88 seconds to build instance. [ 716.255104] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.412031] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.447238] env[62368]: INFO nova.compute.manager [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] [instance: 70055136-b08b-4654-aca7-8280158d94da] Took 1.03 seconds to deallocate network for instance. [ 716.664860] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec83400a-deab-45e7-bf01-99d194adee39 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.674431] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b317944f-dd71-4f29-bf88-63132905b95c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.711108] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135def3f-332d-4edf-8237-c9a216e480a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.718565] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704925df-b922-4de1-ae5e-92e75ed40b99 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.737450] env[62368]: DEBUG nova.compute.provider_tree [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.749675] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.754792] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77895342-972b-4c46-afa4-167af8b19941 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.696s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.916661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Releasing lock "refresh_cache-5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.916915] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 716.917111] env[62368]: DEBUG nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.917280] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 716.936053] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.242197] env[62368]: DEBUG nova.scheduler.client.report [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.264386] env[62368]: DEBUG nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.436837] env[62368]: DEBUG nova.network.neutron [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.486015] env[62368]: INFO nova.scheduler.client.report [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Deleted allocations for instance 70055136-b08b-4654-aca7-8280158d94da [ 717.750462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.039s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.751231] env[62368]: ERROR nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Traceback (most recent call last): [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self.driver.spawn(context, instance, image_meta, [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] vm_ref = self.build_virtual_machine(instance, [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.751231] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] for vif in network_info: [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] return self._sync_wrapper(fn, *args, **kwargs) [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self.wait() [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self[:] = self._gt.wait() [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] return self._exit_event.wait() [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] current.throw(*self._exc) [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.751630] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] result = function(*args, **kwargs) [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] return func(*args, **kwargs) [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] raise e [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] nwinfo = self.network_api.allocate_for_instance( [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] created_port_ids = self._update_ports_for_instance( [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] with excutils.save_and_reraise_exception(): [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] self.force_reraise() [ 717.752117] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] raise self.value [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] updated_port = self._update_port( [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] _ensure_no_port_binding_failure(port) [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] raise exception.PortBindingFailed(port_id=port['id']) [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] nova.exception.PortBindingFailed: Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. [ 717.752519] env[62368]: ERROR nova.compute.manager [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] [ 717.752519] env[62368]: DEBUG nova.compute.utils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.754021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.573s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.755844] env[62368]: INFO nova.compute.claims [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.757607] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Build of instance a7c65d2c-49c1-4c0c-b38b-4e1a89552b67 was re-scheduled: Binding failed for port 29095761-1f84-4074-a398-6db40c48f31c, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 717.758049] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 717.758272] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Acquiring lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.758421] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Acquired lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.758942] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.768045] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquiring lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.768370] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.768623] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquiring lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.768848] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.770891] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.002s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.773561] env[62368]: INFO nova.compute.manager [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Terminating instance [ 717.780998] env[62368]: DEBUG nova.compute.manager [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 717.781514] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 717.782762] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ea8b10-5899-4e66-b269-5ac682337200 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.794601] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 717.796100] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b90b241-aa44-4b3f-bd35-fd0c755a8ff9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.804941] env[62368]: DEBUG oslo_vmware.api [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 717.804941] env[62368]: value = "task-1198200" [ 717.804941] env[62368]: _type = "Task" [ 717.804941] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.816684] env[62368]: DEBUG oslo_vmware.api [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198200, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.818474] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.940312] env[62368]: INFO nova.compute.manager [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] [instance: 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6] Took 1.02 seconds to deallocate network for instance. [ 717.995823] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a37e36b9-4000-4a06-b35e-6d2e6492611b tempest-ServersTestManualDisk-630102089 tempest-ServersTestManualDisk-630102089-project-member] Lock "70055136-b08b-4654-aca7-8280158d94da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.897s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.295951] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.317558] env[62368]: DEBUG oslo_vmware.api [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198200, 'name': PowerOffVM_Task, 'duration_secs': 0.247617} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.317558] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 718.317558] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 718.317949] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bca43218-4fcb-43b0-b31e-d4dfbdacfd42 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.401282] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 718.401581] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 718.401764] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Deleting the datastore file [datastore1] 38291806-d0ae-47a5-bc6e-3d0320b2bd50 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 718.402058] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17223f04-b46a-43ea-8052-92da193c0e07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.417113] env[62368]: DEBUG oslo_vmware.api [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for the task: (returnval){ [ 718.417113] env[62368]: value = "task-1198202" [ 718.417113] env[62368]: _type = "Task" [ 718.417113] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.421803] env[62368]: DEBUG oslo_vmware.api [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198202, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.461494] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.497875] env[62368]: DEBUG nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.589483] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 718.589723] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 718.924446] env[62368]: DEBUG oslo_vmware.api [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Task: {'id': task-1198202, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.485697} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.924772] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 718.926436] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 718.926436] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 718.926436] env[62368]: INFO nova.compute.manager [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Took 1.14 seconds to destroy the instance on the hypervisor. [ 718.926436] env[62368]: DEBUG oslo.service.loopingcall [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.926436] env[62368]: DEBUG nova.compute.manager [-] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.926436] env[62368]: DEBUG nova.network.neutron [-] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.964276] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Releasing lock "refresh_cache-a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.964517] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 718.965016] env[62368]: DEBUG nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.965138] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.973770] env[62368]: INFO nova.scheduler.client.report [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Deleted allocations for instance 5d1dfbdc-1b96-400a-bd0a-e898a90bdab6 [ 718.987948] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.033090] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.100400] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.101542] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 719.101542] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Rebuilding the list of instances to heal {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 719.196013] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166c15a2-314a-4866-bd34-d28bc6adea57 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.205424] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6911659-3916-48e0-8c38-55e94b04fafa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.239095] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62233f0-d83e-4fce-9965-54df1fef4e18 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.248880] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d799e2-71a1-4721-99ce-fc715535e6a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.265649] env[62368]: DEBUG nova.compute.provider_tree [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.321303] env[62368]: DEBUG nova.compute.manager [req-2008540a-dce7-4758-b06c-dc8a88bd95dd req-c871b9aa-cce2-4072-a3d0-1cd21e81d9f4 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Received event network-vif-deleted-2c8ae58a-cd9a-4b06-941a-7863e1d3855f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.321486] env[62368]: INFO nova.compute.manager [req-2008540a-dce7-4758-b06c-dc8a88bd95dd req-c871b9aa-cce2-4072-a3d0-1cd21e81d9f4 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Neutron deleted interface 2c8ae58a-cd9a-4b06-941a-7863e1d3855f; detaching it from the instance and deleting it from the info cache [ 719.321611] env[62368]: DEBUG nova.network.neutron [req-2008540a-dce7-4758-b06c-dc8a88bd95dd req-c871b9aa-cce2-4072-a3d0-1cd21e81d9f4 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.486623] env[62368]: DEBUG oslo_concurrency.lockutils [None req-04a56473-72a1-4dd1-9da5-23171eba076f tempest-ServersTestJSON-1579406837 tempest-ServersTestJSON-1579406837-project-member] Lock "5d1dfbdc-1b96-400a-bd0a-e898a90bdab6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.776s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.490577] env[62368]: DEBUG nova.network.neutron [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.604279] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 719.604279] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Skipping network cache update for instance because it is being deleted. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 719.604493] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 719.604630] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Didn't find any instances for network info cache update. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 719.604759] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.604864] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.605020] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.605167] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.605354] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.605518] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.605643] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 719.605780] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.770407] env[62368]: DEBUG nova.scheduler.client.report [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.790485] env[62368]: DEBUG nova.network.neutron [-] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.823672] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bff017b-69cc-4722-90d9-ceb69e7a072b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.836019] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf821e2-175b-4618-943a-e4ed3e64bfa4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.863099] env[62368]: DEBUG nova.compute.manager [req-2008540a-dce7-4758-b06c-dc8a88bd95dd req-c871b9aa-cce2-4072-a3d0-1cd21e81d9f4 service nova] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Detach interface failed, port_id=2c8ae58a-cd9a-4b06-941a-7863e1d3855f, reason: Instance 38291806-d0ae-47a5-bc6e-3d0320b2bd50 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 719.989685] env[62368]: DEBUG nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.993778] env[62368]: INFO nova.compute.manager [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] [instance: a7c65d2c-49c1-4c0c-b38b-4e1a89552b67] Took 1.03 seconds to deallocate network for instance. [ 720.110781] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.278444] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.278989] env[62368]: DEBUG nova.compute.manager [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 720.281608] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.255s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.293454] env[62368]: INFO nova.compute.manager [-] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Took 1.37 seconds to deallocate network for instance. [ 720.515418] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.786127] env[62368]: DEBUG nova.compute.utils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.790626] env[62368]: DEBUG nova.compute.manager [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Not allocating networking since 'none' was specified. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 720.806833] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.044958] env[62368]: INFO nova.scheduler.client.report [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Deleted allocations for instance a7c65d2c-49c1-4c0c-b38b-4e1a89552b67 [ 721.182347] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727ccada-4aad-40af-9869-7878ba804461 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.191601] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4edca46-239a-451e-b026-6ae0581473ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.228554] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66ecee5-79ff-43e4-8919-66f1f4ae4df5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.236810] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10611d37-5723-43e1-8c59-a773c0a5d115 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.251493] env[62368]: DEBUG nova.compute.provider_tree [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.291083] env[62368]: DEBUG nova.compute.manager [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.560967] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b296dd0-b84e-4467-884b-d70b44369442 tempest-ServerActionsTestJSON-27121796 tempest-ServerActionsTestJSON-27121796-project-member] Lock "a7c65d2c-49c1-4c0c-b38b-4e1a89552b67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.178s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.755592] env[62368]: DEBUG nova.scheduler.client.report [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.065189] env[62368]: DEBUG nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.260959] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.261628] env[62368]: ERROR nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Traceback (most recent call last): [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self.driver.spawn(context, instance, image_meta, [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] vm_ref = self.build_virtual_machine(instance, [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.261628] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] for vif in network_info: [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] return self._sync_wrapper(fn, *args, **kwargs) [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self.wait() [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self[:] = self._gt.wait() [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] return self._exit_event.wait() [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] current.throw(*self._exc) [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.262019] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] result = function(*args, **kwargs) [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] return func(*args, **kwargs) [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] raise e [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] nwinfo = self.network_api.allocate_for_instance( [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] created_port_ids = self._update_ports_for_instance( [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] with excutils.save_and_reraise_exception(): [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] self.force_reraise() [ 722.262381] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] raise self.value [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] updated_port = self._update_port( [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] _ensure_no_port_binding_failure(port) [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] raise exception.PortBindingFailed(port_id=port['id']) [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] nova.exception.PortBindingFailed: Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. [ 722.262724] env[62368]: ERROR nova.compute.manager [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] [ 722.262724] env[62368]: DEBUG nova.compute.utils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 722.263511] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.669s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.265756] env[62368]: INFO nova.compute.claims [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 722.271147] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Build of instance ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf was re-scheduled: Binding failed for port 5604a801-e04f-403a-aed5-44092a88f8ea, please check neutron logs for more information. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 722.272043] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 722.272043] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Acquiring lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.272159] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Acquired lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.272227] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 722.305412] env[62368]: DEBUG nova.compute.manager [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.340430] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.341072] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.341072] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.341072] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.341442] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.341442] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.341515] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.341648] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.341774] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.342843] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.342920] env[62368]: DEBUG nova.virt.hardware [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.343803] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e96c8d7-ce9a-4a5e-a995-a7f6b55737e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.354022] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a666b89-c506-4916-beef-becfa38f0ae0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.369853] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.375974] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Creating folder: Project (3973e582a60748b9933a15c8904b5d56). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 722.376308] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f83f02f-d254-4edc-bb20-27b32b3011a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.387085] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Created folder: Project (3973e582a60748b9933a15c8904b5d56) in parent group-v259706. [ 722.387285] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Creating folder: Instances. Parent ref: group-v259726. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 722.387515] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-add5818e-b03c-4ad8-a4bb-4615e8b95e25 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.396859] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Created folder: Instances in parent group-v259726. [ 722.396859] env[62368]: DEBUG oslo.service.loopingcall [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.396859] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 722.396859] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a06fd663-670c-4063-826e-3af816c3e667 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.413253] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.413253] env[62368]: value = "task-1198205" [ 722.413253] env[62368]: _type = "Task" [ 722.413253] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.420507] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198205, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.595044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.792353] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.914707] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.925951] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198205, 'name': CreateVM_Task, 'duration_secs': 0.266513} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.926724] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 722.927192] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.927387] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.927704] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 722.928175] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d4057c3-dad3-42ee-9ae7-aa9808d78bcb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.932624] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 722.932624] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529c6fa1-b15f-1ba3-0c44-3c954c74d6c6" [ 722.932624] env[62368]: _type = "Task" [ 722.932624] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.941124] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529c6fa1-b15f-1ba3-0c44-3c954c74d6c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.420571] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Releasing lock "refresh_cache-ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.420821] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 723.421236] env[62368]: DEBUG nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.421432] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 723.438602] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.443684] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529c6fa1-b15f-1ba3-0c44-3c954c74d6c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010422} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.449015] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.449015] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.449015] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.449015] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.449404] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.449404] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83fb7884-9656-45c1-889e-20025b62e1de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.459135] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.459313] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 723.460047] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e90b2b35-3054-4128-8f0b-778a0c0f061e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.470360] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 723.470360] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5256f0e2-2c07-e369-989c-2ecd474ad5a5" [ 723.470360] env[62368]: _type = "Task" [ 723.470360] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.479238] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5256f0e2-2c07-e369-989c-2ecd474ad5a5, 'name': SearchDatastore_Task, 'duration_secs': 0.008234} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.480086] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae9cafb3-13cc-42e6-b6f2-79b5b6b86614 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.485741] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 723.485741] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52092882-0837-6f33-605e-645dfc56839b" [ 723.485741] env[62368]: _type = "Task" [ 723.485741] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.497690] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52092882-0837-6f33-605e-645dfc56839b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.633295] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4816595b-9f23-42b0-86d2-8b4a9ce9c43d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.641107] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8669ad69-dd78-4748-89d3-9f1602a0ac41 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.683041] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c9e49a-2e95-484e-a7e5-67b34b1b39e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.692181] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986e874b-0527-4b5d-8c0d-466e46a0aa8c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.707815] env[62368]: DEBUG nova.compute.provider_tree [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.945143] env[62368]: DEBUG nova.network.neutron [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.999187] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52092882-0837-6f33-605e-645dfc56839b, 'name': SearchDatastore_Task, 'duration_secs': 0.032902} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.999453] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.999714] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] a5c638d8-f761-497d-bde2-39ea90d8d151/a5c638d8-f761-497d-bde2-39ea90d8d151.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 723.999966] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-baaa4caf-512c-4b23-b076-59149e8b6c8f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.005929] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 724.005929] env[62368]: value = "task-1198206" [ 724.005929] env[62368]: _type = "Task" [ 724.005929] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.014460] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198206, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.214302] env[62368]: DEBUG nova.scheduler.client.report [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.449847] env[62368]: INFO nova.compute.manager [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] [instance: ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf] Took 1.03 seconds to deallocate network for instance. [ 724.515484] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198206, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475767} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.516422] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] a5c638d8-f761-497d-bde2-39ea90d8d151/a5c638d8-f761-497d-bde2-39ea90d8d151.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 724.516631] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 724.516874] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-948e454f-4745-4d76-b17a-157d5eccda42 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.524187] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 724.524187] env[62368]: value = "task-1198207" [ 724.524187] env[62368]: _type = "Task" [ 724.524187] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.533849] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198207, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.723361] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.724323] env[62368]: DEBUG nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.726768] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.524s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.728188] env[62368]: INFO nova.compute.claims [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.035654] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198207, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059395} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.035844] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 725.036892] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7dd7505-6414-4021-a9f6-4fa373bab408 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.058706] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] a5c638d8-f761-497d-bde2-39ea90d8d151/a5c638d8-f761-497d-bde2-39ea90d8d151.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 725.058875] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28509e58-e2e6-4d77-a63d-a3daa41bddc2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.081090] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 725.081090] env[62368]: value = "task-1198208" [ 725.081090] env[62368]: _type = "Task" [ 725.081090] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.089414] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198208, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.233393] env[62368]: DEBUG nova.compute.utils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.238350] env[62368]: DEBUG nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 725.238612] env[62368]: DEBUG nova.network.neutron [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 725.293683] env[62368]: DEBUG nova.policy [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4329e27dab00428da3ac68f220818db2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '31833b73a8724021bab7a6efedc7acfd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 725.491280] env[62368]: INFO nova.scheduler.client.report [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Deleted allocations for instance ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf [ 725.596025] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198208, 'name': ReconfigVM_Task, 'duration_secs': 0.326738} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.596025] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Reconfigured VM instance instance-0000002a to attach disk [datastore1] a5c638d8-f761-497d-bde2-39ea90d8d151/a5c638d8-f761-497d-bde2-39ea90d8d151.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 725.596025] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e339f2f-46a2-431f-9206-5ac216c92e9a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.602938] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 725.602938] env[62368]: value = "task-1198209" [ 725.602938] env[62368]: _type = "Task" [ 725.602938] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.611661] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198209, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.734802] env[62368]: DEBUG nova.network.neutron [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Successfully created port: d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.739220] env[62368]: DEBUG nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.917974] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.918366] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.002962] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ddaaa9a1-608a-4690-b460-1e1e26bc8c36 tempest-ServerPasswordTestJSON-2132032845 tempest-ServerPasswordTestJSON-2132032845-project-member] Lock "ccdfcfa9-fb1e-4a08-a9fd-2c4bd9837caf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.607s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.078497] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47f725a-31b9-45bb-aae0-72b20d6b6fd6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.088137] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff25ad2-85db-4251-86e1-9b44122f94eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.128337] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a0db5f-0e8a-4b12-a9e1-7bdbbb305df4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.135604] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198209, 'name': Rename_Task, 'duration_secs': 0.144899} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.137683] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 726.138036] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26feda00-e708-4ab0-b488-593b1698f9c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.140702] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7535ee7-4bf5-4986-a040-e03d33f205ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.154544] env[62368]: DEBUG nova.compute.provider_tree [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.156590] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 726.156590] env[62368]: value = "task-1198210" [ 726.156590] env[62368]: _type = "Task" [ 726.156590] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.164318] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198210, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.507138] env[62368]: DEBUG nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.664549] env[62368]: DEBUG nova.scheduler.client.report [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.674076] env[62368]: DEBUG oslo_vmware.api [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198210, 'name': PowerOnVM_Task, 'duration_secs': 0.446302} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.674441] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 726.674846] env[62368]: INFO nova.compute.manager [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Took 4.37 seconds to spawn the instance on the hypervisor. [ 726.675052] env[62368]: DEBUG nova.compute.manager [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 726.675898] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ab3fdb-7787-4256-878e-270f0afa6a4b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.755869] env[62368]: DEBUG nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 726.792299] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.792693] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.792969] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.793431] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.793880] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.794545] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.794545] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.794680] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.794770] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.794931] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.795118] env[62368]: DEBUG nova.virt.hardware [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.796006] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df323a55-0bf4-4e94-8739-aec50a654153 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.806488] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7411a7eb-e9a9-4d76-816a-55bb7341956d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.036048] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.170312] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.170834] env[62368]: DEBUG nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.173449] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.017s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.179062] env[62368]: INFO nova.compute.claims [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.199487] env[62368]: INFO nova.compute.manager [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Took 23.03 seconds to build instance. [ 727.684818] env[62368]: DEBUG nova.compute.utils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.694952] env[62368]: DEBUG nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 727.694952] env[62368]: DEBUG nova.network.neutron [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 727.703580] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f674772d-b54e-4a88-a91e-43c5cd662f65 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "a5c638d8-f761-497d-bde2-39ea90d8d151" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.598s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.709732] env[62368]: DEBUG nova.compute.manager [req-e3607267-a72c-4861-be7f-236c21a8c0dc req-0900c73a-d056-470a-a711-a6e41ea9849c service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Received event network-vif-plugged-d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.709732] env[62368]: DEBUG oslo_concurrency.lockutils [req-e3607267-a72c-4861-be7f-236c21a8c0dc req-0900c73a-d056-470a-a711-a6e41ea9849c service nova] Acquiring lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.709809] env[62368]: DEBUG oslo_concurrency.lockutils [req-e3607267-a72c-4861-be7f-236c21a8c0dc req-0900c73a-d056-470a-a711-a6e41ea9849c service nova] Lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.710046] env[62368]: DEBUG oslo_concurrency.lockutils [req-e3607267-a72c-4861-be7f-236c21a8c0dc req-0900c73a-d056-470a-a711-a6e41ea9849c service nova] Lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.710153] env[62368]: DEBUG nova.compute.manager [req-e3607267-a72c-4861-be7f-236c21a8c0dc req-0900c73a-d056-470a-a711-a6e41ea9849c service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] No waiting events found dispatching network-vif-plugged-d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 727.710316] env[62368]: WARNING nova.compute.manager [req-e3607267-a72c-4861-be7f-236c21a8c0dc req-0900c73a-d056-470a-a711-a6e41ea9849c service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Received unexpected event network-vif-plugged-d98e888b-22dd-483e-999f-4e91c0e15ab6 for instance with vm_state building and task_state spawning. [ 727.778657] env[62368]: DEBUG nova.policy [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96a6cbe0353c44209751845b09e0cbf9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a7e415d782c4b9794232f25b9ae0f20', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.851108] env[62368]: DEBUG nova.network.neutron [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Successfully updated port: d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 728.188714] env[62368]: DEBUG nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.217426] env[62368]: DEBUG nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.252075] env[62368]: DEBUG nova.compute.manager [None req-6a16bd8a-4f9d-4174-bab8-c80778b0d727 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 728.252991] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d2b4c7-89a7-4871-9be8-ac7baba6b55f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.330495] env[62368]: DEBUG nova.network.neutron [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Successfully created port: a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.354266] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.354389] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquired lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.354583] env[62368]: DEBUG nova.network.neutron [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 728.416366] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquiring lock "a5c638d8-f761-497d-bde2-39ea90d8d151" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.418253] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "a5c638d8-f761-497d-bde2-39ea90d8d151" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.419030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquiring lock "a5c638d8-f761-497d-bde2-39ea90d8d151-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.419030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "a5c638d8-f761-497d-bde2-39ea90d8d151-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.419030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "a5c638d8-f761-497d-bde2-39ea90d8d151-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.425018] env[62368]: INFO nova.compute.manager [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Terminating instance [ 728.426260] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquiring lock "refresh_cache-a5c638d8-f761-497d-bde2-39ea90d8d151" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.426535] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquired lock "refresh_cache-a5c638d8-f761-497d-bde2-39ea90d8d151" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.426775] env[62368]: DEBUG nova.network.neutron [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 728.577733] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61470a25-d9e3-49f1-b6d0-7c63bf0a9d19 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.584132] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4652e59-ca02-43f9-b2a0-aeda63946886 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.615643] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c1ec88-fb93-4f3f-9d36-51920e1905d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.624159] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313e8155-0fae-4776-b705-9cefde6217b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.643426] env[62368]: DEBUG nova.compute.provider_tree [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.748995] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.768908] env[62368]: INFO nova.compute.manager [None req-6a16bd8a-4f9d-4174-bab8-c80778b0d727 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] instance snapshotting [ 728.769556] env[62368]: DEBUG nova.objects.instance [None req-6a16bd8a-4f9d-4174-bab8-c80778b0d727 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lazy-loading 'flavor' on Instance uuid a5c638d8-f761-497d-bde2-39ea90d8d151 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 728.903117] env[62368]: DEBUG nova.network.neutron [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.036732] env[62368]: DEBUG nova.network.neutron [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.142419] env[62368]: DEBUG nova.network.neutron [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.144437] env[62368]: DEBUG nova.network.neutron [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.151591] env[62368]: DEBUG nova.scheduler.client.report [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.202390] env[62368]: DEBUG nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.249121] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.249365] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.249519] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.249697] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.249842] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.249986] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.250310] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.250368] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.250525] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.250688] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.250860] env[62368]: DEBUG nova.virt.hardware [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.251788] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411f1536-ba0d-4e2c-bd59-dcb502f32642 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.261914] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ff8e48-6a5e-4d83-95a5-78e7f5980f36 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.278597] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48761c0-b5f0-45a0-a652-d9a7ff4d92a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.295765] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbfbcea1-459b-438e-a8da-501a7d17e466 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.650097] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Releasing lock "refresh_cache-a5c638d8-f761-497d-bde2-39ea90d8d151" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.650097] env[62368]: DEBUG nova.compute.manager [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.650097] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 729.650097] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Releasing lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.650275] env[62368]: DEBUG nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Instance network_info: |[{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 729.651459] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943fdbac-f4c5-4add-a706-0916a5bcc491 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.654906] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:10:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd98e888b-22dd-483e-999f-4e91c0e15ab6', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.665447] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Creating folder: Project (31833b73a8724021bab7a6efedc7acfd). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 729.666332] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.666830] env[62368]: DEBUG nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.669712] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dcf4a7b8-914c-4ff7-a5ea-527179d63798 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.673414] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.923s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.673784] env[62368]: INFO nova.compute.claims [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.679667] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 729.679667] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d78c87c-b662-4dc4-b5b1-5178d7780fc2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.687313] env[62368]: DEBUG oslo_vmware.api [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 729.687313] env[62368]: value = "task-1198212" [ 729.687313] env[62368]: _type = "Task" [ 729.687313] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.690981] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Created folder: Project (31833b73a8724021bab7a6efedc7acfd) in parent group-v259706. [ 729.691219] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Creating folder: Instances. Parent ref: group-v259729. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 729.691973] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f8c0066-e2a8-4152-8eef-77e63c5639fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.697230] env[62368]: DEBUG oslo_vmware.api [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198212, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.704871] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Created folder: Instances in parent group-v259729. [ 729.704871] env[62368]: DEBUG oslo.service.loopingcall [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.704871] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 729.704871] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73e4f4b0-c684-4d27-a23e-96d78a347b2d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.725279] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.725279] env[62368]: value = "task-1198214" [ 729.725279] env[62368]: _type = "Task" [ 729.725279] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.733592] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198214, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.778145] env[62368]: DEBUG nova.compute.manager [req-aa351f0b-7627-4ed2-a74a-c90e234d6f88 req-1afa0865-475b-46cd-ba4a-cab506fcbdf2 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Received event network-changed-d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.778473] env[62368]: DEBUG nova.compute.manager [req-aa351f0b-7627-4ed2-a74a-c90e234d6f88 req-1afa0865-475b-46cd-ba4a-cab506fcbdf2 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Refreshing instance network info cache due to event network-changed-d98e888b-22dd-483e-999f-4e91c0e15ab6. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.778473] env[62368]: DEBUG oslo_concurrency.lockutils [req-aa351f0b-7627-4ed2-a74a-c90e234d6f88 req-1afa0865-475b-46cd-ba4a-cab506fcbdf2 service nova] Acquiring lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.778609] env[62368]: DEBUG oslo_concurrency.lockutils [req-aa351f0b-7627-4ed2-a74a-c90e234d6f88 req-1afa0865-475b-46cd-ba4a-cab506fcbdf2 service nova] Acquired lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.778714] env[62368]: DEBUG nova.network.neutron [req-aa351f0b-7627-4ed2-a74a-c90e234d6f88 req-1afa0865-475b-46cd-ba4a-cab506fcbdf2 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Refreshing network info cache for port d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 729.809727] env[62368]: DEBUG nova.compute.manager [None req-6a16bd8a-4f9d-4174-bab8-c80778b0d727 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Instance disappeared during snapshot {{(pid=62368) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 730.013110] env[62368]: DEBUG nova.compute.manager [None req-6a16bd8a-4f9d-4174-bab8-c80778b0d727 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Found 0 images (rotation: 2) {{(pid=62368) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 730.174374] env[62368]: DEBUG nova.compute.utils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.175886] env[62368]: DEBUG nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.176099] env[62368]: DEBUG nova.network.neutron [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 730.198668] env[62368]: DEBUG oslo_vmware.api [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198212, 'name': PowerOffVM_Task, 'duration_secs': 0.135322} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.198973] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 730.199158] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 730.199546] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd89c1dd-9bf4-4da8-b01d-c2087ee8aeb7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.225350] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 730.225461] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 730.225614] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Deleting the datastore file [datastore1] a5c638d8-f761-497d-bde2-39ea90d8d151 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 730.225919] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b524128c-1624-45c3-b303-aca3f39047d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.237052] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198214, 'name': CreateVM_Task, 'duration_secs': 0.329181} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.239327] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 730.239327] env[62368]: DEBUG oslo_vmware.api [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for the task: (returnval){ [ 730.239327] env[62368]: value = "task-1198216" [ 730.239327] env[62368]: _type = "Task" [ 730.239327] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.239327] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.239327] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.239600] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 730.239908] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6aa6e527-8a43-45b3-899d-4a1869498445 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.247821] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 730.247821] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c2570f-8e69-41a4-e544-7be5ebe76ef2" [ 730.247821] env[62368]: _type = "Task" [ 730.247821] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.260282] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c2570f-8e69-41a4-e544-7be5ebe76ef2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.266123] env[62368]: DEBUG nova.policy [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6972fdf35fc4fe9afa65cdfd0d46568', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb560320635f493fbe4308af1623e8ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.498157] env[62368]: DEBUG nova.compute.manager [req-cd7f27b9-e4b8-4b56-9fd8-ba18a16428d5 req-ed46c4c5-a23f-41b1-9e43-1d55ffbde68f service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Received event network-vif-plugged-a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.498157] env[62368]: DEBUG oslo_concurrency.lockutils [req-cd7f27b9-e4b8-4b56-9fd8-ba18a16428d5 req-ed46c4c5-a23f-41b1-9e43-1d55ffbde68f service nova] Acquiring lock "3f36000e-b93d-4dda-ac39-b8459203c227-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.498157] env[62368]: DEBUG oslo_concurrency.lockutils [req-cd7f27b9-e4b8-4b56-9fd8-ba18a16428d5 req-ed46c4c5-a23f-41b1-9e43-1d55ffbde68f service nova] Lock "3f36000e-b93d-4dda-ac39-b8459203c227-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.498157] env[62368]: DEBUG oslo_concurrency.lockutils [req-cd7f27b9-e4b8-4b56-9fd8-ba18a16428d5 req-ed46c4c5-a23f-41b1-9e43-1d55ffbde68f service nova] Lock "3f36000e-b93d-4dda-ac39-b8459203c227-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.498157] env[62368]: DEBUG nova.compute.manager [req-cd7f27b9-e4b8-4b56-9fd8-ba18a16428d5 req-ed46c4c5-a23f-41b1-9e43-1d55ffbde68f service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] No waiting events found dispatching network-vif-plugged-a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 730.498636] env[62368]: WARNING nova.compute.manager [req-cd7f27b9-e4b8-4b56-9fd8-ba18a16428d5 req-ed46c4c5-a23f-41b1-9e43-1d55ffbde68f service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Received unexpected event network-vif-plugged-a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0 for instance with vm_state building and task_state spawning. [ 730.593165] env[62368]: DEBUG nova.network.neutron [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Successfully created port: 4ca48f55-7ae9-49a3-917d-cb0a6eec5043 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.657252] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.657490] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.682512] env[62368]: DEBUG nova.network.neutron [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Successfully updated port: a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 730.682512] env[62368]: DEBUG nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.714296] env[62368]: DEBUG nova.network.neutron [req-aa351f0b-7627-4ed2-a74a-c90e234d6f88 req-1afa0865-475b-46cd-ba4a-cab506fcbdf2 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updated VIF entry in instance network info cache for port d98e888b-22dd-483e-999f-4e91c0e15ab6. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 730.714651] env[62368]: DEBUG nova.network.neutron [req-aa351f0b-7627-4ed2-a74a-c90e234d6f88 req-1afa0865-475b-46cd-ba4a-cab506fcbdf2 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.752267] env[62368]: DEBUG oslo_vmware.api [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Task: {'id': task-1198216, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123938} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.755675] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 730.755894] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 730.756100] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 730.756311] env[62368]: INFO nova.compute.manager [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Took 1.11 seconds to destroy the instance on the hypervisor. [ 730.756512] env[62368]: DEBUG oslo.service.loopingcall [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.758979] env[62368]: DEBUG nova.compute.manager [-] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.759093] env[62368]: DEBUG nova.network.neutron [-] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 730.765851] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c2570f-8e69-41a4-e544-7be5ebe76ef2, 'name': SearchDatastore_Task, 'duration_secs': 0.014645} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.766146] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.766370] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 730.766592] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.766737] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.767365] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 730.767622] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5465f5ec-3a9d-45e3-9852-017449386c41 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.775571] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 730.775760] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 730.776570] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85190fdc-8e29-4f53-bc44-b3eaae6592a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.782528] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 730.782528] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524bc8e3-8567-8c1e-a67a-e231b71f0688" [ 730.782528] env[62368]: _type = "Task" [ 730.782528] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.783171] env[62368]: DEBUG nova.network.neutron [-] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.796678] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524bc8e3-8567-8c1e-a67a-e231b71f0688, 'name': SearchDatastore_Task, 'duration_secs': 0.008556} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.797873] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51b7f0c6-60f0-40e1-864e-868cc96b2874 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.805332] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 730.805332] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526aa76c-d0b4-a861-f0f4-a362d791cb52" [ 730.805332] env[62368]: _type = "Task" [ 730.805332] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.813150] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526aa76c-d0b4-a861-f0f4-a362d791cb52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.011840] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4665074d-2f86-4345-a6e9-0e023bc2fba0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.019428] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e1f91a-9bf6-4ffc-b359-bf81bc118c04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.053625] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bde6660-0787-4b58-b127-3e79247eec3c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.061850] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d5824a-e3d0-4ec4-b393-ab95452c8b2d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.076642] env[62368]: DEBUG nova.compute.provider_tree [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 731.197034] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquiring lock "refresh_cache-3f36000e-b93d-4dda-ac39-b8459203c227" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.197146] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquired lock "refresh_cache-3f36000e-b93d-4dda-ac39-b8459203c227" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.197292] env[62368]: DEBUG nova.network.neutron [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 731.219720] env[62368]: DEBUG oslo_concurrency.lockutils [req-aa351f0b-7627-4ed2-a74a-c90e234d6f88 req-1afa0865-475b-46cd-ba4a-cab506fcbdf2 service nova] Releasing lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.287837] env[62368]: DEBUG nova.network.neutron [-] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.316503] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526aa76c-d0b4-a861-f0f4-a362d791cb52, 'name': SearchDatastore_Task, 'duration_secs': 0.008326} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.316731] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.316973] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072/8b7ad7ed-8f2d-4c76-b335-4a2ce5327072.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 731.317221] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e98b30e-2e61-4b6b-8423-1dec87fe626e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.324377] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 731.324377] env[62368]: value = "task-1198217" [ 731.324377] env[62368]: _type = "Task" [ 731.324377] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.331965] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.599232] env[62368]: ERROR nova.scheduler.client.report [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [req-f389f2ed-6749-4c7e-b88c-5814b9da5133] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f389f2ed-6749-4c7e-b88c-5814b9da5133"}]} [ 731.616827] env[62368]: DEBUG nova.scheduler.client.report [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 731.636651] env[62368]: DEBUG nova.scheduler.client.report [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 731.636768] env[62368]: DEBUG nova.compute.provider_tree [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 731.653961] env[62368]: DEBUG nova.scheduler.client.report [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 731.678901] env[62368]: DEBUG nova.scheduler.client.report [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 731.703024] env[62368]: DEBUG nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.732817] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.733115] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.733313] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.733506] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.733719] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.733904] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.734226] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.734346] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.734473] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.734695] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.734889] env[62368]: DEBUG nova.virt.hardware [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.735853] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702c33fa-a0d4-40d3-96ad-088f2582a318 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.741909] env[62368]: DEBUG nova.network.neutron [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.750188] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4de0f78-8c92-44b6-bec8-ee37b036b491 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.790339] env[62368]: INFO nova.compute.manager [-] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Took 1.03 seconds to deallocate network for instance. [ 731.836073] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198217, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.941608] env[62368]: DEBUG nova.network.neutron [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Updating instance_info_cache with network_info: [{"id": "a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0", "address": "fa:16:3e:09:00:40", "network": {"id": "27544ba2-2c00-4241-85cb-36dd18d0cf6b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-638498757-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8a7e415d782c4b9794232f25b9ae0f20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa31cb003-b2", "ovs_interfaceid": "a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.012673] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1a673c-c650-42d2-b57c-bd48b747e27d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.020212] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b9e983-baea-4828-8a29-0ef91cdf0b4f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.051263] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fbe440-d0d8-410d-b436-63aa5e534bef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.058716] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21b245a-ae11-4ec7-8613-285604e311f4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.074039] env[62368]: DEBUG nova.compute.provider_tree [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 732.279748] env[62368]: DEBUG nova.compute.manager [req-4ebbc882-7b28-4b09-b1de-3da956c0de88 req-ff4509c4-5b36-48c3-9425-4d2987bb21c2 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Received event network-vif-plugged-4ca48f55-7ae9-49a3-917d-cb0a6eec5043 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.279909] env[62368]: DEBUG oslo_concurrency.lockutils [req-4ebbc882-7b28-4b09-b1de-3da956c0de88 req-ff4509c4-5b36-48c3-9425-4d2987bb21c2 service nova] Acquiring lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.280207] env[62368]: DEBUG oslo_concurrency.lockutils [req-4ebbc882-7b28-4b09-b1de-3da956c0de88 req-ff4509c4-5b36-48c3-9425-4d2987bb21c2 service nova] Lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.280301] env[62368]: DEBUG oslo_concurrency.lockutils [req-4ebbc882-7b28-4b09-b1de-3da956c0de88 req-ff4509c4-5b36-48c3-9425-4d2987bb21c2 service nova] Lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.281017] env[62368]: DEBUG nova.compute.manager [req-4ebbc882-7b28-4b09-b1de-3da956c0de88 req-ff4509c4-5b36-48c3-9425-4d2987bb21c2 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] No waiting events found dispatching network-vif-plugged-4ca48f55-7ae9-49a3-917d-cb0a6eec5043 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 732.281017] env[62368]: WARNING nova.compute.manager [req-4ebbc882-7b28-4b09-b1de-3da956c0de88 req-ff4509c4-5b36-48c3-9425-4d2987bb21c2 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Received unexpected event network-vif-plugged-4ca48f55-7ae9-49a3-917d-cb0a6eec5043 for instance with vm_state building and task_state spawning. [ 732.300318] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.334481] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585962} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.334722] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072/8b7ad7ed-8f2d-4c76-b335-4a2ce5327072.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 732.334925] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.335182] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df4b1d15-ae16-495c-8031-e34487bf6ec7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.341607] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 732.341607] env[62368]: value = "task-1198218" [ 732.341607] env[62368]: _type = "Task" [ 732.341607] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.348866] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198218, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.372989] env[62368]: DEBUG nova.network.neutron [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Successfully updated port: 4ca48f55-7ae9-49a3-917d-cb0a6eec5043 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 732.443888] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Releasing lock "refresh_cache-3f36000e-b93d-4dda-ac39-b8459203c227" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.444350] env[62368]: DEBUG nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Instance network_info: |[{"id": "a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0", "address": "fa:16:3e:09:00:40", "network": {"id": "27544ba2-2c00-4241-85cb-36dd18d0cf6b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-638498757-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8a7e415d782c4b9794232f25b9ae0f20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa31cb003-b2", "ovs_interfaceid": "a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 732.444796] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:00:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8614db14-cc04-466b-b309-367ab8296cda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 732.452257] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Creating folder: Project (8a7e415d782c4b9794232f25b9ae0f20). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 732.452811] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ddd7b4b-80d8-43a9-8d53-6b87201de260 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.463263] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Created folder: Project (8a7e415d782c4b9794232f25b9ae0f20) in parent group-v259706. [ 732.463448] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Creating folder: Instances. Parent ref: group-v259732. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 732.463670] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-13266ab8-1720-46d7-88a3-86130857796a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.473204] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Created folder: Instances in parent group-v259732. [ 732.473445] env[62368]: DEBUG oslo.service.loopingcall [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.473630] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 732.473820] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-804059a1-ea73-4025-8a61-71d852e5dbaf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.492812] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 732.492812] env[62368]: value = "task-1198221" [ 732.492812] env[62368]: _type = "Task" [ 732.492812] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.500127] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198221, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.609545] env[62368]: DEBUG nova.scheduler.client.report [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 72 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 732.609814] env[62368]: DEBUG nova.compute.provider_tree [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 72 to 73 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 732.609994] env[62368]: DEBUG nova.compute.provider_tree [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 732.632249] env[62368]: DEBUG nova.compute.manager [req-fe6b3d4c-8857-4a30-a154-77201d9c1024 req-708ccff1-6f48-4494-b6e3-65368614c124 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Received event network-changed-a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.632357] env[62368]: DEBUG nova.compute.manager [req-fe6b3d4c-8857-4a30-a154-77201d9c1024 req-708ccff1-6f48-4494-b6e3-65368614c124 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Refreshing instance network info cache due to event network-changed-a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 732.632605] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe6b3d4c-8857-4a30-a154-77201d9c1024 req-708ccff1-6f48-4494-b6e3-65368614c124 service nova] Acquiring lock "refresh_cache-3f36000e-b93d-4dda-ac39-b8459203c227" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.632726] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe6b3d4c-8857-4a30-a154-77201d9c1024 req-708ccff1-6f48-4494-b6e3-65368614c124 service nova] Acquired lock "refresh_cache-3f36000e-b93d-4dda-ac39-b8459203c227" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.632918] env[62368]: DEBUG nova.network.neutron [req-fe6b3d4c-8857-4a30-a154-77201d9c1024 req-708ccff1-6f48-4494-b6e3-65368614c124 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Refreshing network info cache for port a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.852178] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198218, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073246} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.853917] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.853917] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5c2d0c-0341-489b-86d4-f04b2d874052 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.874866] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072/8b7ad7ed-8f2d-4c76-b335-4a2ce5327072.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.875178] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e656297-9ed5-4d97-81a7-6f9aee30c2bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.890478] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "refresh_cache-e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.890626] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "refresh_cache-e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.890775] env[62368]: DEBUG nova.network.neutron [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.898650] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 732.898650] env[62368]: value = "task-1198222" [ 732.898650] env[62368]: _type = "Task" [ 732.898650] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.911011] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198222, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.003881] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198221, 'name': CreateVM_Task, 'duration_secs': 0.389336} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.004080] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 733.004803] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.004967] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.005288] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 733.005534] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75d0ebb6-1a2f-4cd0-8a7a-b99a54cefb88 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.009792] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 733.009792] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fbe542-b5d3-5cd8-13e9-5e291e092a0b" [ 733.009792] env[62368]: _type = "Task" [ 733.009792] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.017130] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fbe542-b5d3-5cd8-13e9-5e291e092a0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.114899] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.443s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.116217] env[62368]: DEBUG nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.118194] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.300s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.119626] env[62368]: INFO nova.compute.claims [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.341474] env[62368]: DEBUG nova.network.neutron [req-fe6b3d4c-8857-4a30-a154-77201d9c1024 req-708ccff1-6f48-4494-b6e3-65368614c124 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Updated VIF entry in instance network info cache for port a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 733.341819] env[62368]: DEBUG nova.network.neutron [req-fe6b3d4c-8857-4a30-a154-77201d9c1024 req-708ccff1-6f48-4494-b6e3-65368614c124 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Updating instance_info_cache with network_info: [{"id": "a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0", "address": "fa:16:3e:09:00:40", "network": {"id": "27544ba2-2c00-4241-85cb-36dd18d0cf6b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-638498757-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8a7e415d782c4b9794232f25b9ae0f20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8614db14-cc04-466b-b309-367ab8296cda", "external-id": "nsx-vlan-transportzone-337", "segmentation_id": 337, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa31cb003-b2", "ovs_interfaceid": "a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.409375] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198222, 'name': ReconfigVM_Task, 'duration_secs': 0.274094} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.409720] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072/8b7ad7ed-8f2d-4c76-b335-4a2ce5327072.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 733.410792] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39451a40-9356-43b9-a099-e1c38bb18519 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.417853] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 733.417853] env[62368]: value = "task-1198223" [ 733.417853] env[62368]: _type = "Task" [ 733.417853] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.421151] env[62368]: DEBUG nova.network.neutron [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.428126] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198223, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.522413] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fbe542-b5d3-5cd8-13e9-5e291e092a0b, 'name': SearchDatastore_Task, 'duration_secs': 0.008797} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.522711] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.522952] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 733.523239] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.523393] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.523573] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 733.523882] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ec1a35f-cebb-43ca-b14a-27d6419ea743 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.531520] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 733.531700] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 733.532446] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-859bd4f5-ae0a-4223-9f0a-8102c4156d11 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.537541] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 733.537541] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527e507e-e159-941c-3e9a-51de90020ba7" [ 733.537541] env[62368]: _type = "Task" [ 733.537541] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.545161] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527e507e-e159-941c-3e9a-51de90020ba7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.552129] env[62368]: DEBUG nova.network.neutron [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Updating instance_info_cache with network_info: [{"id": "4ca48f55-7ae9-49a3-917d-cb0a6eec5043", "address": "fa:16:3e:38:86:63", "network": {"id": "e89b0140-71be-4766-9ec4-c37da482a0e3", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2110723737-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb560320635f493fbe4308af1623e8ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ca48f55-7a", "ovs_interfaceid": "4ca48f55-7ae9-49a3-917d-cb0a6eec5043", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.627146] env[62368]: DEBUG nova.compute.utils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.628784] env[62368]: DEBUG nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.628881] env[62368]: DEBUG nova.network.neutron [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 733.693970] env[62368]: DEBUG nova.policy [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '83b960e3cb1747cca6b5d30b362d7c5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b4bbe6b942c40d3a046937ef8367370', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.845206] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe6b3d4c-8857-4a30-a154-77201d9c1024 req-708ccff1-6f48-4494-b6e3-65368614c124 service nova] Releasing lock "refresh_cache-3f36000e-b93d-4dda-ac39-b8459203c227" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.929456] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198223, 'name': Rename_Task, 'duration_secs': 0.128914} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.932084] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 733.932084] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-517a6df4-c679-4561-b042-1ac85b03f495 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.937899] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 733.937899] env[62368]: value = "task-1198224" [ 733.937899] env[62368]: _type = "Task" [ 733.937899] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.947249] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198224, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.011589] env[62368]: DEBUG nova.network.neutron [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Successfully created port: fe5fd04f-da09-4430-91aa-a990e3a3f4df {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.049137] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527e507e-e159-941c-3e9a-51de90020ba7, 'name': SearchDatastore_Task, 'duration_secs': 0.007863} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.050144] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0b489e6-f5ce-48f0-8ec6-f49fec336327 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.054589] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "refresh_cache-e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.054961] env[62368]: DEBUG nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Instance network_info: |[{"id": "4ca48f55-7ae9-49a3-917d-cb0a6eec5043", "address": "fa:16:3e:38:86:63", "network": {"id": "e89b0140-71be-4766-9ec4-c37da482a0e3", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2110723737-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb560320635f493fbe4308af1623e8ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ca48f55-7a", "ovs_interfaceid": "4ca48f55-7ae9-49a3-917d-cb0a6eec5043", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 734.056767] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:86:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f44b2fa3-6730-4b87-8839-947eff21213f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ca48f55-7ae9-49a3-917d-cb0a6eec5043', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 734.066025] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Creating folder: Project (fb560320635f493fbe4308af1623e8ca). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 734.066520] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 734.066520] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526f4b68-1258-a5e6-50ea-b9687527ec4d" [ 734.066520] env[62368]: _type = "Task" [ 734.066520] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.066665] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dbf158ce-f9d8-4998-930f-fe03dc031b59 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.078555] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526f4b68-1258-a5e6-50ea-b9687527ec4d, 'name': SearchDatastore_Task, 'duration_secs': 0.009743} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.079256] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.079256] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3f36000e-b93d-4dda-ac39-b8459203c227/3f36000e-b93d-4dda-ac39-b8459203c227.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 734.079256] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e513c5d-2200-4eb3-aca5-a2f97c905cea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.081873] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Created folder: Project (fb560320635f493fbe4308af1623e8ca) in parent group-v259706. [ 734.082057] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Creating folder: Instances. Parent ref: group-v259735. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 734.082535] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a06db77c-5d9e-4adf-8a47-d0ae65b9fd80 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.086053] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 734.086053] env[62368]: value = "task-1198226" [ 734.086053] env[62368]: _type = "Task" [ 734.086053] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.091073] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Created folder: Instances in parent group-v259735. [ 734.091271] env[62368]: DEBUG oslo.service.loopingcall [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.094250] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 734.094511] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198226, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.094728] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b46b34e8-5705-41b0-a60c-e2602d99db05 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.113172] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 734.113172] env[62368]: value = "task-1198228" [ 734.113172] env[62368]: _type = "Task" [ 734.113172] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.121138] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198228, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.132698] env[62368]: DEBUG nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.436465] env[62368]: DEBUG nova.compute.manager [req-26981e71-686f-4b88-ae30-bb8bba0cd1a4 req-63c7e5c6-2f73-4ae2-ba30-f8a8edb09d70 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Received event network-changed-4ca48f55-7ae9-49a3-917d-cb0a6eec5043 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.436703] env[62368]: DEBUG nova.compute.manager [req-26981e71-686f-4b88-ae30-bb8bba0cd1a4 req-63c7e5c6-2f73-4ae2-ba30-f8a8edb09d70 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Refreshing instance network info cache due to event network-changed-4ca48f55-7ae9-49a3-917d-cb0a6eec5043. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 734.436880] env[62368]: DEBUG oslo_concurrency.lockutils [req-26981e71-686f-4b88-ae30-bb8bba0cd1a4 req-63c7e5c6-2f73-4ae2-ba30-f8a8edb09d70 service nova] Acquiring lock "refresh_cache-e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.437024] env[62368]: DEBUG oslo_concurrency.lockutils [req-26981e71-686f-4b88-ae30-bb8bba0cd1a4 req-63c7e5c6-2f73-4ae2-ba30-f8a8edb09d70 service nova] Acquired lock "refresh_cache-e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.437418] env[62368]: DEBUG nova.network.neutron [req-26981e71-686f-4b88-ae30-bb8bba0cd1a4 req-63c7e5c6-2f73-4ae2-ba30-f8a8edb09d70 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Refreshing network info cache for port 4ca48f55-7ae9-49a3-917d-cb0a6eec5043 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.453981] env[62368]: DEBUG oslo_vmware.api [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198224, 'name': PowerOnVM_Task, 'duration_secs': 0.477791} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.456953] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 734.457186] env[62368]: INFO nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Took 7.70 seconds to spawn the instance on the hypervisor. [ 734.457405] env[62368]: DEBUG nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 734.458844] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a942da4-4f97-4bca-ac59-6f27a92c70b5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.510325] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a61d95-4c23-4b28-8a74-baf7462ac076 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.519306] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3faebca9-daf2-49ce-a5e8-eebe677302cd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.552729] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855c41bb-7898-42f1-a899-c4b90eec51f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.560620] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b34bbf-3332-42f5-a25d-2ffce96d1196 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.575237] env[62368]: DEBUG nova.compute.provider_tree [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.594940] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198226, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490385} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.595187] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3f36000e-b93d-4dda-ac39-b8459203c227/3f36000e-b93d-4dda-ac39-b8459203c227.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 734.595397] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.595626] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1676ed8-8e29-4da1-9930-229a3dd071e0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.600700] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 734.600700] env[62368]: value = "task-1198229" [ 734.600700] env[62368]: _type = "Task" [ 734.600700] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.607723] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198229, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.621424] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198228, 'name': CreateVM_Task, 'duration_secs': 0.45567} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.621579] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 734.622260] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.622435] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.622758] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 734.623008] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47e39dd1-376c-4008-ab59-e40204ba72a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.626917] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 734.626917] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f82a9c-29e3-b20f-df59-1b6b8af4d2aa" [ 734.626917] env[62368]: _type = "Task" [ 734.626917] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.634228] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f82a9c-29e3-b20f-df59-1b6b8af4d2aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.980978] env[62368]: INFO nova.compute.manager [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Took 26.41 seconds to build instance. [ 735.077859] env[62368]: DEBUG nova.scheduler.client.report [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.110288] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198229, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062532} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.112900] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.113691] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36497d1-9440-4f8b-b4d5-8b765efa0f25 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.137494] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 3f36000e-b93d-4dda-ac39-b8459203c227/3f36000e-b93d-4dda-ac39-b8459203c227.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.141137] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7118a80c-3d08-4328-90ef-439f0de1062a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.159049] env[62368]: DEBUG nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.167957] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f82a9c-29e3-b20f-df59-1b6b8af4d2aa, 'name': SearchDatastore_Task, 'duration_secs': 0.009396} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.171495] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.171736] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 735.171971] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.172129] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.172309] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 735.172845] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 735.172845] env[62368]: value = "task-1198230" [ 735.172845] env[62368]: _type = "Task" [ 735.172845] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.173142] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a52cd95-ff2d-47a6-944c-409595493ca4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.186268] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.195217] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.195481] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.195636] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.195812] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.195953] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.196801] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.197069] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.197234] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.197405] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.197566] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.197856] env[62368]: DEBUG nova.virt.hardware [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.198679] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5336f58-abcc-41f9-bcb9-3691ca23ecab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.203539] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 735.203712] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 735.206391] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a100fe7b-9694-4e5d-8beb-3a0fe7d45f18 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.210056] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf59999d-39a9-4258-9321-6bc57f109fe7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.216404] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 735.216404] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f82fa3-d3ea-5924-4595-245e4c5f82b6" [ 735.216404] env[62368]: _type = "Task" [ 735.216404] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.227984] env[62368]: DEBUG nova.network.neutron [req-26981e71-686f-4b88-ae30-bb8bba0cd1a4 req-63c7e5c6-2f73-4ae2-ba30-f8a8edb09d70 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Updated VIF entry in instance network info cache for port 4ca48f55-7ae9-49a3-917d-cb0a6eec5043. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 735.227984] env[62368]: DEBUG nova.network.neutron [req-26981e71-686f-4b88-ae30-bb8bba0cd1a4 req-63c7e5c6-2f73-4ae2-ba30-f8a8edb09d70 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Updating instance_info_cache with network_info: [{"id": "4ca48f55-7ae9-49a3-917d-cb0a6eec5043", "address": "fa:16:3e:38:86:63", "network": {"id": "e89b0140-71be-4766-9ec4-c37da482a0e3", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2110723737-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb560320635f493fbe4308af1623e8ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ca48f55-7a", "ovs_interfaceid": "4ca48f55-7ae9-49a3-917d-cb0a6eec5043", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.234800] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f82fa3-d3ea-5924-4595-245e4c5f82b6, 'name': SearchDatastore_Task, 'duration_secs': 0.009594} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.235566] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d6e6852-fa5d-470f-ba27-30778eb64a94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.241040] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 735.241040] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5229b579-65a7-1cac-c3d8-b3978eafd681" [ 735.241040] env[62368]: _type = "Task" [ 735.241040] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.249736] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5229b579-65a7-1cac-c3d8-b3978eafd681, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.483404] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47cb62c3-f663-47e5-87e7-978b641c044f tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.376s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.582587] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.583159] env[62368]: DEBUG nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.588013] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.553s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.588013] env[62368]: INFO nova.compute.claims [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.685744] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198230, 'name': ReconfigVM_Task, 'duration_secs': 0.330628} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.686087] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 3f36000e-b93d-4dda-ac39-b8459203c227/3f36000e-b93d-4dda-ac39-b8459203c227.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 735.686708] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3178d6d4-5b87-4054-a5b0-7595b73c5333 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.693197] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 735.693197] env[62368]: value = "task-1198231" [ 735.693197] env[62368]: _type = "Task" [ 735.693197] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.703168] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198231, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.730480] env[62368]: DEBUG oslo_concurrency.lockutils [req-26981e71-686f-4b88-ae30-bb8bba0cd1a4 req-63c7e5c6-2f73-4ae2-ba30-f8a8edb09d70 service nova] Releasing lock "refresh_cache-e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.750745] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5229b579-65a7-1cac-c3d8-b3978eafd681, 'name': SearchDatastore_Task, 'duration_secs': 0.011874} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.750997] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.751279] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0/e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 735.751541] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84139648-0042-4126-ac1d-5c4c56a2d6eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.757804] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 735.757804] env[62368]: value = "task-1198232" [ 735.757804] env[62368]: _type = "Task" [ 735.757804] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.764738] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198232, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.807680] env[62368]: DEBUG nova.network.neutron [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Successfully updated port: fe5fd04f-da09-4430-91aa-a990e3a3f4df {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 735.987532] env[62368]: DEBUG nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.091389] env[62368]: DEBUG nova.compute.utils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.095321] env[62368]: DEBUG nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.095612] env[62368]: DEBUG nova.network.neutron [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 736.184806] env[62368]: DEBUG nova.policy [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77af709b6fad45889037a8f0bf7169a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd6a19a2bf826456398e46c2df1b805e0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.204547] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198231, 'name': Rename_Task, 'duration_secs': 0.164006} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.205054] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 736.206831] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3874a7b1-8014-4cdb-b9aa-bd7e79d54ae3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.215955] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 736.215955] env[62368]: value = "task-1198233" [ 736.215955] env[62368]: _type = "Task" [ 736.215955] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.228944] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.270865] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198232, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5016} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.271145] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0/e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 736.271383] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 736.271630] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f580186d-7398-4fcb-8e48-47633d0d6090 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.277618] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 736.277618] env[62368]: value = "task-1198234" [ 736.277618] env[62368]: _type = "Task" [ 736.277618] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.286728] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198234, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.312880] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "refresh_cache-4dc7b6b4-1bf5-4195-bb93-14756f8f9986" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.312880] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquired lock "refresh_cache-4dc7b6b4-1bf5-4195-bb93-14756f8f9986" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.312880] env[62368]: DEBUG nova.network.neutron [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.469426] env[62368]: DEBUG nova.compute.manager [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Received event network-changed-d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.469608] env[62368]: DEBUG nova.compute.manager [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Refreshing instance network info cache due to event network-changed-d98e888b-22dd-483e-999f-4e91c0e15ab6. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 736.469828] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Acquiring lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.470596] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Acquired lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.470832] env[62368]: DEBUG nova.network.neutron [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Refreshing network info cache for port d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 736.509818] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.596790] env[62368]: DEBUG nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.701162] env[62368]: DEBUG nova.network.neutron [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Successfully created port: 6db52e60-d5f1-4028-b5ff-ffd3bc376cfb {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.725066] env[62368]: DEBUG oslo_vmware.api [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198233, 'name': PowerOnVM_Task, 'duration_secs': 0.497361} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.725361] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 736.725518] env[62368]: INFO nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Took 7.52 seconds to spawn the instance on the hypervisor. [ 736.725754] env[62368]: DEBUG nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 736.726505] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9973ddf8-f0e3-4fbd-a344-3da1a78f5055 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.788304] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198234, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063948} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.788780] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 736.789327] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3158c8d3-ce71-45d8-903d-1510221c95bc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.829205] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0/e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 736.840025] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3832973-3ba9-402c-afd1-5b3f1441939b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.870372] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 736.870372] env[62368]: value = "task-1198235" [ 736.870372] env[62368]: _type = "Task" [ 736.870372] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.887866] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198235, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.907443] env[62368]: DEBUG nova.network.neutron [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.039300] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc77da6-bb91-481a-9bbf-b7e2ded15e1e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.046933] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0526340-561c-4d26-b8d7-c24ddb0adecd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.083529] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d3f49a-0dfd-476f-8aa5-964512a2b151 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.091876] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067fb5ea-fc03-4128-a5da-2fa53e2aff91 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.112390] env[62368]: DEBUG nova.compute.provider_tree [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 737.226299] env[62368]: DEBUG nova.network.neutron [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Updating instance_info_cache with network_info: [{"id": "fe5fd04f-da09-4430-91aa-a990e3a3f4df", "address": "fa:16:3e:6c:c6:4b", "network": {"id": "c50b1235-56db-4af7-bac8-8867341b31f7", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-643724583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b4bbe6b942c40d3a046937ef8367370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "095fbf26-7367-4f9e-87c5-2965b64b0b0f", "external-id": "nsx-vlan-transportzone-777", "segmentation_id": 777, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe5fd04f-da", "ovs_interfaceid": "fe5fd04f-da09-4430-91aa-a990e3a3f4df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.245020] env[62368]: INFO nova.compute.manager [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Took 24.05 seconds to build instance. [ 737.361315] env[62368]: DEBUG nova.network.neutron [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updated VIF entry in instance network info cache for port d98e888b-22dd-483e-999f-4e91c0e15ab6. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 737.361877] env[62368]: DEBUG nova.network.neutron [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.381787] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198235, 'name': ReconfigVM_Task, 'duration_secs': 0.431854} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.382090] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Reconfigured VM instance instance-0000002d to attach disk [datastore1] e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0/e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 737.382746] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbd816cc-ee52-4303-8494-d3f929072281 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.389273] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 737.389273] env[62368]: value = "task-1198236" [ 737.389273] env[62368]: _type = "Task" [ 737.389273] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.401500] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198236, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.615798] env[62368]: DEBUG nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.657677] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.657925] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.658089] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.658271] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.658417] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.658559] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.659261] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.659261] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.659261] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.659261] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.659455] env[62368]: DEBUG nova.virt.hardware [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.661045] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6bc81f-98b9-42be-93ef-9045cf8b7031 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.664690] env[62368]: DEBUG nova.scheduler.client.report [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 73 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 737.664690] env[62368]: DEBUG nova.compute.provider_tree [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 73 to 74 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 737.664855] env[62368]: DEBUG nova.compute.provider_tree [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 737.671433] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6cb66b-506d-4bb8-8eb8-b21d6fa87647 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.729240] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Releasing lock "refresh_cache-4dc7b6b4-1bf5-4195-bb93-14756f8f9986" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.729581] env[62368]: DEBUG nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Instance network_info: |[{"id": "fe5fd04f-da09-4430-91aa-a990e3a3f4df", "address": "fa:16:3e:6c:c6:4b", "network": {"id": "c50b1235-56db-4af7-bac8-8867341b31f7", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-643724583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b4bbe6b942c40d3a046937ef8367370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "095fbf26-7367-4f9e-87c5-2965b64b0b0f", "external-id": "nsx-vlan-transportzone-777", "segmentation_id": 777, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe5fd04f-da", "ovs_interfaceid": "fe5fd04f-da09-4430-91aa-a990e3a3f4df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 737.730031] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:c6:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '095fbf26-7367-4f9e-87c5-2965b64b0b0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe5fd04f-da09-4430-91aa-a990e3a3f4df', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.738279] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Creating folder: Project (3b4bbe6b942c40d3a046937ef8367370). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.738616] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-394521d3-0162-4411-aeee-cc42e5adafe8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.747208] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9caa98b2-74c4-4aed-9fc5-1832de993ada tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "3f36000e-b93d-4dda-ac39-b8459203c227" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.278s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.749737] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Created folder: Project (3b4bbe6b942c40d3a046937ef8367370) in parent group-v259706. [ 737.749914] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Creating folder: Instances. Parent ref: group-v259738. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.750357] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8534cc62-d578-444b-a8b1-687c328a816d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.760895] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Created folder: Instances in parent group-v259738. [ 737.761198] env[62368]: DEBUG oslo.service.loopingcall [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.761470] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 737.761771] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ac247f3-9bbc-4e35-9f8f-2e849bb70e84 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.783008] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.783008] env[62368]: value = "task-1198239" [ 737.783008] env[62368]: _type = "Task" [ 737.783008] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.790043] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198239, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.864426] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Releasing lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.864730] env[62368]: DEBUG nova.compute.manager [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Received event network-vif-plugged-fe5fd04f-da09-4430-91aa-a990e3a3f4df {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.864936] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Acquiring lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.865159] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.865351] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.865521] env[62368]: DEBUG nova.compute.manager [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] No waiting events found dispatching network-vif-plugged-fe5fd04f-da09-4430-91aa-a990e3a3f4df {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 737.865766] env[62368]: WARNING nova.compute.manager [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Received unexpected event network-vif-plugged-fe5fd04f-da09-4430-91aa-a990e3a3f4df for instance with vm_state building and task_state spawning. [ 737.865955] env[62368]: DEBUG nova.compute.manager [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Received event network-changed-fe5fd04f-da09-4430-91aa-a990e3a3f4df {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.866140] env[62368]: DEBUG nova.compute.manager [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Refreshing instance network info cache due to event network-changed-fe5fd04f-da09-4430-91aa-a990e3a3f4df. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 737.866333] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Acquiring lock "refresh_cache-4dc7b6b4-1bf5-4195-bb93-14756f8f9986" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.866468] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Acquired lock "refresh_cache-4dc7b6b4-1bf5-4195-bb93-14756f8f9986" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.866621] env[62368]: DEBUG nova.network.neutron [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Refreshing network info cache for port fe5fd04f-da09-4430-91aa-a990e3a3f4df {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 737.898779] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198236, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.171184] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.171713] env[62368]: DEBUG nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.178582] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.065s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.178771] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.005s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.178932] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 738.179256] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.665s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.180703] env[62368]: INFO nova.compute.claims [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.185381] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9761a7a-36f7-43ed-aa8a-aed4e4a03dab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.193991] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349be2b0-402c-472d-bc60-c029d9770b86 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.211481] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acefedbf-6c52-4436-9b82-a3321fc65ca9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.222038] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd8b09a-9c3a-4bc5-8c9f-c819c884d295 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.256280] env[62368]: DEBUG nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.260102] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181362MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 738.260102] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.261371] env[62368]: DEBUG nova.network.neutron [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Successfully updated port: 6db52e60-d5f1-4028-b5ff-ffd3bc376cfb {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 738.290982] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198239, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.399913] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198236, 'name': Rename_Task, 'duration_secs': 0.559046} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.399971] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 738.400252] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4ac2c8c-8224-426a-861f-49a1da8368a4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.407651] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 738.407651] env[62368]: value = "task-1198240" [ 738.407651] env[62368]: _type = "Task" [ 738.407651] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.415963] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198240, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.575710] env[62368]: DEBUG nova.compute.manager [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Received event network-vif-plugged-6db52e60-d5f1-4028-b5ff-ffd3bc376cfb {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.575926] env[62368]: DEBUG oslo_concurrency.lockutils [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] Acquiring lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.576141] env[62368]: DEBUG oslo_concurrency.lockutils [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] Lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.576306] env[62368]: DEBUG oslo_concurrency.lockutils [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] Lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.576479] env[62368]: DEBUG nova.compute.manager [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] No waiting events found dispatching network-vif-plugged-6db52e60-d5f1-4028-b5ff-ffd3bc376cfb {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 738.576640] env[62368]: WARNING nova.compute.manager [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Received unexpected event network-vif-plugged-6db52e60-d5f1-4028-b5ff-ffd3bc376cfb for instance with vm_state building and task_state spawning. [ 738.576794] env[62368]: DEBUG nova.compute.manager [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Received event network-changed-6db52e60-d5f1-4028-b5ff-ffd3bc376cfb {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.576943] env[62368]: DEBUG nova.compute.manager [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Refreshing instance network info cache due to event network-changed-6db52e60-d5f1-4028-b5ff-ffd3bc376cfb. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.577536] env[62368]: DEBUG oslo_concurrency.lockutils [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] Acquiring lock "refresh_cache-9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.577591] env[62368]: DEBUG oslo_concurrency.lockutils [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] Acquired lock "refresh_cache-9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.577737] env[62368]: DEBUG nova.network.neutron [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Refreshing network info cache for port 6db52e60-d5f1-4028-b5ff-ffd3bc376cfb {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 738.655491] env[62368]: DEBUG nova.network.neutron [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Updated VIF entry in instance network info cache for port fe5fd04f-da09-4430-91aa-a990e3a3f4df. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 738.655830] env[62368]: DEBUG nova.network.neutron [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Updating instance_info_cache with network_info: [{"id": "fe5fd04f-da09-4430-91aa-a990e3a3f4df", "address": "fa:16:3e:6c:c6:4b", "network": {"id": "c50b1235-56db-4af7-bac8-8867341b31f7", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-643724583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b4bbe6b942c40d3a046937ef8367370", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "095fbf26-7367-4f9e-87c5-2965b64b0b0f", "external-id": "nsx-vlan-transportzone-777", "segmentation_id": 777, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe5fd04f-da", "ovs_interfaceid": "fe5fd04f-da09-4430-91aa-a990e3a3f4df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.686108] env[62368]: DEBUG nova.compute.utils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.691476] env[62368]: DEBUG nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.691476] env[62368]: DEBUG nova.network.neutron [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 738.739867] env[62368]: DEBUG nova.policy [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5f3d8a68af04196b6bf358a87f78bc1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a05bbb0b9b104e70b562f8abaa26a164', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.765931] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquiring lock "refresh_cache-9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.783321] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.796872] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198239, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.918610] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198240, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.921553] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquiring lock "3f36000e-b93d-4dda-ac39-b8459203c227" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.921788] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "3f36000e-b93d-4dda-ac39-b8459203c227" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.922053] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquiring lock "3f36000e-b93d-4dda-ac39-b8459203c227-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.922173] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "3f36000e-b93d-4dda-ac39-b8459203c227-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.922343] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "3f36000e-b93d-4dda-ac39-b8459203c227-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.925140] env[62368]: INFO nova.compute.manager [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Terminating instance [ 738.929023] env[62368]: DEBUG nova.compute.manager [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 738.929023] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 738.929023] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25bff51c-f77f-41e2-8ad9-164210de4194 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.934839] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 738.935069] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48019908-5221-4e04-ac10-93d22404bd1c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.941389] env[62368]: DEBUG oslo_vmware.api [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 738.941389] env[62368]: value = "task-1198241" [ 738.941389] env[62368]: _type = "Task" [ 738.941389] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.949060] env[62368]: DEBUG oslo_vmware.api [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198241, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.080475] env[62368]: DEBUG nova.network.neutron [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Successfully created port: 4b2bdc5d-6176-4415-a60e-9519b40bd055 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.158916] env[62368]: DEBUG oslo_concurrency.lockutils [req-35dcd77b-d046-452f-9521-130461115129 req-08bfea88-fad0-40be-9989-4b703801dc9b service nova] Releasing lock "refresh_cache-4dc7b6b4-1bf5-4195-bb93-14756f8f9986" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.191941] env[62368]: DEBUG nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.296969] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198239, 'name': CreateVM_Task, 'duration_secs': 1.09096} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.296969] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 739.297697] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.297825] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.298163] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 739.298993] env[62368]: DEBUG nova.network.neutron [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.301193] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1df7a23b-adfa-4893-810d-b25f6f23b93b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.307292] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 739.307292] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5296781f-4b0e-16ee-94b8-bffecd258765" [ 739.307292] env[62368]: _type = "Task" [ 739.307292] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.321185] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5296781f-4b0e-16ee-94b8-bffecd258765, 'name': SearchDatastore_Task, 'duration_secs': 0.010259} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.321567] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.321841] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 739.322171] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.322298] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.322517] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 739.322802] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75841d75-1f46-44b3-a18c-fcad89b0e99d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.331019] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 739.331187] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 739.331884] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2eeaebb9-157f-43ac-a312-957b6def418e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.337524] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 739.337524] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520ad843-bcec-657f-69cb-dce7f662ba5c" [ 739.337524] env[62368]: _type = "Task" [ 739.337524] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.351589] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520ad843-bcec-657f-69cb-dce7f662ba5c, 'name': SearchDatastore_Task, 'duration_secs': 0.008244} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.354262] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e13d8da-a91b-4a3c-bcc3-17854387220a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.359354] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 739.359354] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521d71a4-fadd-f6fe-5723-f91398c0250f" [ 739.359354] env[62368]: _type = "Task" [ 739.359354] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.367493] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521d71a4-fadd-f6fe-5723-f91398c0250f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.413503] env[62368]: DEBUG nova.network.neutron [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.422607] env[62368]: DEBUG oslo_vmware.api [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198240, 'name': PowerOnVM_Task, 'duration_secs': 0.768358} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.422856] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 739.423303] env[62368]: INFO nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Took 7.72 seconds to spawn the instance on the hypervisor. [ 739.423527] env[62368]: DEBUG nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 739.424370] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d26913-a3b4-4a4c-8603-ddf1dec040df {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.452688] env[62368]: DEBUG oslo_vmware.api [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198241, 'name': PowerOffVM_Task, 'duration_secs': 0.144326} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.452688] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 739.452688] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 739.452688] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e7de2ba-2238-41d0-8d45-915d2dba22fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.512945] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 739.513245] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 739.513474] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Deleting the datastore file [datastore2] 3f36000e-b93d-4dda-ac39-b8459203c227 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 739.513752] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0642a139-89da-42db-a553-aa6b0da280f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.519990] env[62368]: DEBUG oslo_vmware.api [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for the task: (returnval){ [ 739.519990] env[62368]: value = "task-1198243" [ 739.519990] env[62368]: _type = "Task" [ 739.519990] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.530288] env[62368]: DEBUG oslo_vmware.api [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.566439] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb0d681-58ba-41f3-8d25-66e8226cd439 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.573503] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d79fa72-59d1-474e-aa2d-a365d36578ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.606714] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1876e97-d1a2-4cbe-8940-d20e0553b8bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.615265] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaaac472-c451-4b96-af90-ce1531afae0f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.630849] env[62368]: DEBUG nova.compute.provider_tree [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.872823] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521d71a4-fadd-f6fe-5723-f91398c0250f, 'name': SearchDatastore_Task, 'duration_secs': 0.009501} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.872823] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.872823] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 4dc7b6b4-1bf5-4195-bb93-14756f8f9986/4dc7b6b4-1bf5-4195-bb93-14756f8f9986.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 739.872823] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1145b721-c213-4161-a7a9-eb824cf14d91 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.879385] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 739.879385] env[62368]: value = "task-1198244" [ 739.879385] env[62368]: _type = "Task" [ 739.879385] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.887262] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198244, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.916102] env[62368]: DEBUG oslo_concurrency.lockutils [req-4a044080-11b5-461d-9687-870c25c99cc3 req-3989211f-7831-48f6-bf8c-c8a2d3eb633f service nova] Releasing lock "refresh_cache-9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.916433] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquired lock "refresh_cache-9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.916595] env[62368]: DEBUG nova.network.neutron [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.945021] env[62368]: INFO nova.compute.manager [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Took 24.83 seconds to build instance. [ 740.033311] env[62368]: DEBUG oslo_vmware.api [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Task: {'id': task-1198243, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151554} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.033613] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 740.033805] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 740.033985] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 740.034192] env[62368]: INFO nova.compute.manager [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Took 1.11 seconds to destroy the instance on the hypervisor. [ 740.034502] env[62368]: DEBUG oslo.service.loopingcall [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.034725] env[62368]: DEBUG nova.compute.manager [-] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.034822] env[62368]: DEBUG nova.network.neutron [-] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 740.137195] env[62368]: DEBUG nova.scheduler.client.report [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.208038] env[62368]: DEBUG nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 740.241879] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.242799] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.243056] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.243342] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.243555] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.243714] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.243972] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.244172] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.244368] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.244566] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.244785] env[62368]: DEBUG nova.virt.hardware [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.245791] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ea68b1-a564-4ce6-b6e9-d017e1c2f48e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.255442] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08062cd-d559-4ac9-b247-a800da8ac806 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.363063] env[62368]: DEBUG nova.compute.manager [req-7c49020d-bc87-4c7e-b781-b0c472d96a57 req-21c0e952-096f-4e16-a417-94160a6ba3e1 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Received event network-vif-deleted-a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.363345] env[62368]: INFO nova.compute.manager [req-7c49020d-bc87-4c7e-b781-b0c472d96a57 req-21c0e952-096f-4e16-a417-94160a6ba3e1 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Neutron deleted interface a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0; detaching it from the instance and deleting it from the info cache [ 740.363521] env[62368]: DEBUG nova.network.neutron [req-7c49020d-bc87-4c7e-b781-b0c472d96a57 req-21c0e952-096f-4e16-a417-94160a6ba3e1 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.386695] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198244, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467451} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.386936] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 4dc7b6b4-1bf5-4195-bb93-14756f8f9986/4dc7b6b4-1bf5-4195-bb93-14756f8f9986.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 740.387255] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 740.387504] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9fc7355-5ea0-4c7d-9e6c-535712306eb7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.393342] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 740.393342] env[62368]: value = "task-1198245" [ 740.393342] env[62368]: _type = "Task" [ 740.393342] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.401144] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198245, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.445168] env[62368]: DEBUG oslo_concurrency.lockutils [None req-669bd994-a224-4efa-9e08-04bf8c482d0d tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.813s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.464418] env[62368]: DEBUG nova.network.neutron [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.642249] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.642879] env[62368]: DEBUG nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.648664] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.840s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.648664] env[62368]: DEBUG nova.objects.instance [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lazy-loading 'resources' on Instance uuid 38291806-d0ae-47a5-bc6e-3d0320b2bd50 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 740.659572] env[62368]: DEBUG nova.network.neutron [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Updating instance_info_cache with network_info: [{"id": "6db52e60-d5f1-4028-b5ff-ffd3bc376cfb", "address": "fa:16:3e:6a:97:ef", "network": {"id": "6ff8eff5-653a-4569-a12d-fae88b453de0", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-973918509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d6a19a2bf826456398e46c2df1b805e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6db52e60-d5", "ovs_interfaceid": "6db52e60-d5f1-4028-b5ff-ffd3bc376cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.693208] env[62368]: DEBUG nova.compute.manager [req-34adb229-12ae-49c1-bbab-5e63a623bf90 req-ddcd696a-6e41-4ac8-a25a-6f8925c004b5 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Received event network-vif-plugged-4b2bdc5d-6176-4415-a60e-9519b40bd055 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.693663] env[62368]: DEBUG oslo_concurrency.lockutils [req-34adb229-12ae-49c1-bbab-5e63a623bf90 req-ddcd696a-6e41-4ac8-a25a-6f8925c004b5 service nova] Acquiring lock "b393f654-0611-493e-a89f-445dfd3f8c13-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.694109] env[62368]: DEBUG oslo_concurrency.lockutils [req-34adb229-12ae-49c1-bbab-5e63a623bf90 req-ddcd696a-6e41-4ac8-a25a-6f8925c004b5 service nova] Lock "b393f654-0611-493e-a89f-445dfd3f8c13-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.694450] env[62368]: DEBUG oslo_concurrency.lockutils [req-34adb229-12ae-49c1-bbab-5e63a623bf90 req-ddcd696a-6e41-4ac8-a25a-6f8925c004b5 service nova] Lock "b393f654-0611-493e-a89f-445dfd3f8c13-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.694782] env[62368]: DEBUG nova.compute.manager [req-34adb229-12ae-49c1-bbab-5e63a623bf90 req-ddcd696a-6e41-4ac8-a25a-6f8925c004b5 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] No waiting events found dispatching network-vif-plugged-4b2bdc5d-6176-4415-a60e-9519b40bd055 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 740.695157] env[62368]: WARNING nova.compute.manager [req-34adb229-12ae-49c1-bbab-5e63a623bf90 req-ddcd696a-6e41-4ac8-a25a-6f8925c004b5 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Received unexpected event network-vif-plugged-4b2bdc5d-6176-4415-a60e-9519b40bd055 for instance with vm_state building and task_state spawning. [ 740.843883] env[62368]: DEBUG nova.network.neutron [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Successfully updated port: 4b2bdc5d-6176-4415-a60e-9519b40bd055 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 740.843883] env[62368]: DEBUG nova.network.neutron [-] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.868213] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33fef912-5226-4f57-8cc0-2dd0b60458a5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.874912] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e876e438-efb1-4241-966a-eaea021b803c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.899197] env[62368]: DEBUG nova.compute.manager [req-7c49020d-bc87-4c7e-b781-b0c472d96a57 req-21c0e952-096f-4e16-a417-94160a6ba3e1 service nova] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Detach interface failed, port_id=a31cb003-b23a-4f5e-a96e-4d5bf46bf6e0, reason: Instance 3f36000e-b93d-4dda-ac39-b8459203c227 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 740.906708] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198245, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058918} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.907640] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.907726] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26f77b1-8b1c-4243-a5f3-f51d698b8d84 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.929466] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 4dc7b6b4-1bf5-4195-bb93-14756f8f9986/4dc7b6b4-1bf5-4195-bb93-14756f8f9986.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.929738] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5798554-c1a2-4a23-a719-0aa7a6c27242 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.947650] env[62368]: DEBUG nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.951366] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 740.951366] env[62368]: value = "task-1198246" [ 740.951366] env[62368]: _type = "Task" [ 740.951366] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.960216] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198246, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.149864] env[62368]: DEBUG nova.compute.utils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.155449] env[62368]: DEBUG nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 741.155449] env[62368]: DEBUG nova.network.neutron [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 741.163065] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Releasing lock "refresh_cache-9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.163065] env[62368]: DEBUG nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Instance network_info: |[{"id": "6db52e60-d5f1-4028-b5ff-ffd3bc376cfb", "address": "fa:16:3e:6a:97:ef", "network": {"id": "6ff8eff5-653a-4569-a12d-fae88b453de0", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-973918509-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d6a19a2bf826456398e46c2df1b805e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6db52e60-d5", "ovs_interfaceid": "6db52e60-d5f1-4028-b5ff-ffd3bc376cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 741.163647] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:97:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6db52e60-d5f1-4028-b5ff-ffd3bc376cfb', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 741.172374] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Creating folder: Project (d6a19a2bf826456398e46c2df1b805e0). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 741.172374] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1c1b246-b095-4832-8478-601458518c4d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.187086] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Created folder: Project (d6a19a2bf826456398e46c2df1b805e0) in parent group-v259706. [ 741.187424] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Creating folder: Instances. Parent ref: group-v259741. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 741.187724] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98e7edf4-9a10-4a2d-b4a2-24ab0daf58a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.200512] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Created folder: Instances in parent group-v259741. [ 741.200912] env[62368]: DEBUG oslo.service.loopingcall [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.201602] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 741.201602] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecf3665f-a117-4713-95de-9b29524ff175 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.218471] env[62368]: DEBUG nova.policy [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '738b6865b4dd4290b8695177a4c66acb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b12e7fa4eec4f3b975651f2fc46c8fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 741.226982] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 741.226982] env[62368]: value = "task-1198249" [ 741.226982] env[62368]: _type = "Task" [ 741.226982] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.237311] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198249, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.345687] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquiring lock "refresh_cache-b393f654-0611-493e-a89f-445dfd3f8c13" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.345870] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquired lock "refresh_cache-b393f654-0611-493e-a89f-445dfd3f8c13" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.346272] env[62368]: DEBUG nova.network.neutron [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.347352] env[62368]: INFO nova.compute.manager [-] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Took 1.31 seconds to deallocate network for instance. [ 741.472365] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198246, 'name': ReconfigVM_Task, 'duration_secs': 0.289336} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.472633] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 4dc7b6b4-1bf5-4195-bb93-14756f8f9986/4dc7b6b4-1bf5-4195-bb93-14756f8f9986.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 741.477274] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5bc06f5b-fe66-40a0-ad1a-8411558f7497 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.481805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.484720] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 741.484720] env[62368]: value = "task-1198250" [ 741.484720] env[62368]: _type = "Task" [ 741.484720] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.493850] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198250, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.502496] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0256c29-30a9-4328-bcfb-740a65c548d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.510520] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e843ef8a-d9ce-410a-b6e3-682e1143ac1f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.548841] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d44df83-4bac-4339-abfa-a0e33d317671 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.555958] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39b42ae-e42b-45f8-85ac-49a11b07454f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.564041] env[62368]: DEBUG nova.network.neutron [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Successfully created port: 3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.574451] env[62368]: DEBUG nova.compute.provider_tree [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.655539] env[62368]: DEBUG nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 741.739690] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198249, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.858319] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.893698] env[62368]: DEBUG nova.network.neutron [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.994693] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198250, 'name': Rename_Task, 'duration_secs': 0.146566} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.996141] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 741.996141] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab6a7192-a780-4bf6-974f-fda7116b8546 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.001985] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 742.001985] env[62368]: value = "task-1198251" [ 742.001985] env[62368]: _type = "Task" [ 742.001985] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.012302] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198251, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.076262] env[62368]: DEBUG nova.scheduler.client.report [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.130067] env[62368]: DEBUG nova.network.neutron [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Updating instance_info_cache with network_info: [{"id": "4b2bdc5d-6176-4415-a60e-9519b40bd055", "address": "fa:16:3e:55:db:a0", "network": {"id": "0cd4e6b5-a19e-41b3-8d4b-8133bf04f6f7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-675197317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a05bbb0b9b104e70b562f8abaa26a164", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b2bdc5d-61", "ovs_interfaceid": "4b2bdc5d-6176-4415-a60e-9519b40bd055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.239464] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198249, 'name': CreateVM_Task, 'duration_secs': 0.516784} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.239764] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 742.240520] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.240699] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.241418] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 742.241532] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fad3312c-12aa-4797-a9bd-e4db34748f5a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.247032] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 742.247032] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5218944e-fc0c-cfd3-e711-1ea637131ad0" [ 742.247032] env[62368]: _type = "Task" [ 742.247032] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.257116] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5218944e-fc0c-cfd3-e711-1ea637131ad0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.513380] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198251, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.581740] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.585110] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.989s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.586222] env[62368]: INFO nova.compute.claims [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.604239] env[62368]: INFO nova.scheduler.client.report [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Deleted allocations for instance 38291806-d0ae-47a5-bc6e-3d0320b2bd50 [ 742.633596] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Releasing lock "refresh_cache-b393f654-0611-493e-a89f-445dfd3f8c13" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.633766] env[62368]: DEBUG nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Instance network_info: |[{"id": "4b2bdc5d-6176-4415-a60e-9519b40bd055", "address": "fa:16:3e:55:db:a0", "network": {"id": "0cd4e6b5-a19e-41b3-8d4b-8133bf04f6f7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-675197317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a05bbb0b9b104e70b562f8abaa26a164", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b2bdc5d-61", "ovs_interfaceid": "4b2bdc5d-6176-4415-a60e-9519b40bd055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 742.634067] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:db:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b29df12-5674-476d-a9e5-5e20f704d224', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b2bdc5d-6176-4415-a60e-9519b40bd055', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 742.641541] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Creating folder: Project (a05bbb0b9b104e70b562f8abaa26a164). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.642183] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8308e497-ee89-4c48-a626-ed602735ddc8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.652935] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Created folder: Project (a05bbb0b9b104e70b562f8abaa26a164) in parent group-v259706. [ 742.653135] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Creating folder: Instances. Parent ref: group-v259744. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.653912] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-199475c0-ab5b-4e20-9e57-823a4003608b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.663983] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Created folder: Instances in parent group-v259744. [ 742.664214] env[62368]: DEBUG oslo.service.loopingcall [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.664395] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 742.664578] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fcb6707d-b076-4299-bf8f-1ac9913df0a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.679054] env[62368]: DEBUG nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.686874] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.686874] env[62368]: value = "task-1198254" [ 742.686874] env[62368]: _type = "Task" [ 742.686874] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.694344] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198254, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.701547] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.701771] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.701923] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.702114] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.702258] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.702403] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.702604] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.702757] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.702916] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.703089] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.703335] env[62368]: DEBUG nova.virt.hardware [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.704023] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1e2bbe-fbee-41db-ae40-38dd8bedf08d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.710759] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d602464-6439-45bc-b35e-4982c013522a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.727339] env[62368]: DEBUG nova.compute.manager [req-1daad25a-b7b3-4d1f-8093-48db86c7442c req-7b7cc76b-1aea-4ac3-91dd-a13af17093d4 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Received event network-changed-4b2bdc5d-6176-4415-a60e-9519b40bd055 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.727587] env[62368]: DEBUG nova.compute.manager [req-1daad25a-b7b3-4d1f-8093-48db86c7442c req-7b7cc76b-1aea-4ac3-91dd-a13af17093d4 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Refreshing instance network info cache due to event network-changed-4b2bdc5d-6176-4415-a60e-9519b40bd055. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 742.727745] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daad25a-b7b3-4d1f-8093-48db86c7442c req-7b7cc76b-1aea-4ac3-91dd-a13af17093d4 service nova] Acquiring lock "refresh_cache-b393f654-0611-493e-a89f-445dfd3f8c13" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.727884] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daad25a-b7b3-4d1f-8093-48db86c7442c req-7b7cc76b-1aea-4ac3-91dd-a13af17093d4 service nova] Acquired lock "refresh_cache-b393f654-0611-493e-a89f-445dfd3f8c13" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.728047] env[62368]: DEBUG nova.network.neutron [req-1daad25a-b7b3-4d1f-8093-48db86c7442c req-7b7cc76b-1aea-4ac3-91dd-a13af17093d4 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Refreshing network info cache for port 4b2bdc5d-6176-4415-a60e-9519b40bd055 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.757636] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5218944e-fc0c-cfd3-e711-1ea637131ad0, 'name': SearchDatastore_Task, 'duration_secs': 0.012122} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.758017] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.758326] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.758617] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.758825] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.759085] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.759395] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ac541e2-f343-47a9-98ff-a3d920c111f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.772011] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.772216] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 742.772969] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71fce254-87e6-4c30-b800-bf26e4fae180 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.778520] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 742.778520] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52989537-e4c2-d9ae-a245-5850d9a4a1c2" [ 742.778520] env[62368]: _type = "Task" [ 742.778520] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.786145] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52989537-e4c2-d9ae-a245-5850d9a4a1c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.829348] env[62368]: DEBUG nova.compute.manager [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 742.830236] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fbac9de-2239-459b-b105-fb0b8c3e1742 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.013146] env[62368]: DEBUG oslo_vmware.api [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198251, 'name': PowerOnVM_Task, 'duration_secs': 0.514648} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.013490] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 743.013682] env[62368]: INFO nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Took 7.85 seconds to spawn the instance on the hypervisor. [ 743.013866] env[62368]: DEBUG nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 743.014624] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4969be0a-5031-42c9-b0be-007b60f58e2d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.113854] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f3e07d2-d2da-419e-99a9-1e7a5624d790 tempest-ServerMetadataNegativeTestJSON-423634474 tempest-ServerMetadataNegativeTestJSON-423634474-project-member] Lock "38291806-d0ae-47a5-bc6e-3d0320b2bd50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.345s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.198487] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198254, 'name': CreateVM_Task, 'duration_secs': 0.356187} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.198487] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 743.198487] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.198487] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.198487] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 743.198677] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19d36463-c5a0-41ef-a0bb-ad51ae99982c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.204107] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 743.204107] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52341d95-6c34-8511-dc00-ea6fcb68f410" [ 743.204107] env[62368]: _type = "Task" [ 743.204107] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.211527] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52341d95-6c34-8511-dc00-ea6fcb68f410, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.289207] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52989537-e4c2-d9ae-a245-5850d9a4a1c2, 'name': SearchDatastore_Task, 'duration_secs': 0.015773} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.290327] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-218ff4a7-f1a3-4969-b1f2-2b9ac269f77c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.295746] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 743.295746] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520d7f0a-0dfd-baf7-4f6e-060c37d7f7ab" [ 743.295746] env[62368]: _type = "Task" [ 743.295746] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.303563] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520d7f0a-0dfd-baf7-4f6e-060c37d7f7ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.336971] env[62368]: DEBUG nova.network.neutron [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Successfully updated port: 3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 743.340866] env[62368]: INFO nova.compute.manager [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] instance snapshotting [ 743.346264] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0027fe9-fbb0-4589-9b12-9a9c190dc6d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.375642] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c02f3a-8691-493f-bf3b-61584cd8ffbe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.536061] env[62368]: INFO nova.compute.manager [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Took 26.80 seconds to build instance. [ 743.648976] env[62368]: DEBUG nova.network.neutron [req-1daad25a-b7b3-4d1f-8093-48db86c7442c req-7b7cc76b-1aea-4ac3-91dd-a13af17093d4 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Updated VIF entry in instance network info cache for port 4b2bdc5d-6176-4415-a60e-9519b40bd055. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 743.648976] env[62368]: DEBUG nova.network.neutron [req-1daad25a-b7b3-4d1f-8093-48db86c7442c req-7b7cc76b-1aea-4ac3-91dd-a13af17093d4 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Updating instance_info_cache with network_info: [{"id": "4b2bdc5d-6176-4415-a60e-9519b40bd055", "address": "fa:16:3e:55:db:a0", "network": {"id": "0cd4e6b5-a19e-41b3-8d4b-8133bf04f6f7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-675197317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a05bbb0b9b104e70b562f8abaa26a164", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b29df12-5674-476d-a9e5-5e20f704d224", "external-id": "nsx-vlan-transportzone-754", "segmentation_id": 754, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b2bdc5d-61", "ovs_interfaceid": "4b2bdc5d-6176-4415-a60e-9519b40bd055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.716397] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52341d95-6c34-8511-dc00-ea6fcb68f410, 'name': SearchDatastore_Task, 'duration_secs': 0.033189} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.716683] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.716910] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 743.717196] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.810455] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520d7f0a-0dfd-baf7-4f6e-060c37d7f7ab, 'name': SearchDatastore_Task, 'duration_secs': 0.020039} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.814161] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.814161] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4/9ce5c5a1-5c90-4f76-a511-d8f855b6bab4.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 743.814161] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.814161] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 743.814421] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f33e1af4-c7c3-49f8-9b7a-aef27dd10dc6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.816598] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1495057-4c66-448b-8718-55bb026224bb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.822834] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 743.822834] env[62368]: value = "task-1198255" [ 743.822834] env[62368]: _type = "Task" [ 743.822834] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.830316] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 743.830482] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 743.831649] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-511ed330-b21e-43ee-8349-eb0b972d98a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.836994] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198255, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.840379] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 743.840379] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e5e827-d7e9-b463-ea81-4dd3266bb26e" [ 743.840379] env[62368]: _type = "Task" [ 743.840379] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.846640] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquiring lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.846640] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquired lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.846798] env[62368]: DEBUG nova.network.neutron [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.852923] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e5e827-d7e9-b463-ea81-4dd3266bb26e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.891295] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 743.892041] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-237c91f6-c09e-4326-9c48-ed8ac781a846 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.900594] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 743.900594] env[62368]: value = "task-1198256" [ 743.900594] env[62368]: _type = "Task" [ 743.900594] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.911714] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198256, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.960999] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d755de5c-3f8d-4639-9bc2-cf2cc6a3cc22 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.968532] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539e900b-66de-4863-9c9a-d4d0f20bf0b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.000453] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4089cf03-a376-4cc3-9507-6e4f20f56d76 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.008407] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7d4477-7c6e-4850-813f-1480c4c5952a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.022767] env[62368]: DEBUG nova.compute.provider_tree [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.037993] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1a8baef9-ee80-44a5-b65f-7ab5a061b04b tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.235s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.152220] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daad25a-b7b3-4d1f-8093-48db86c7442c req-7b7cc76b-1aea-4ac3-91dd-a13af17093d4 service nova] Releasing lock "refresh_cache-b393f654-0611-493e-a89f-445dfd3f8c13" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.334840] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198255, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.353165] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e5e827-d7e9-b463-ea81-4dd3266bb26e, 'name': SearchDatastore_Task, 'duration_secs': 0.017294} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.354548] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-815ac73c-a2f1-4408-8510-b0084f146341 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.366761] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 744.366761] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526387e3-f92d-6fc2-0b05-62b59113d101" [ 744.366761] env[62368]: _type = "Task" [ 744.366761] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.378598] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526387e3-f92d-6fc2-0b05-62b59113d101, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.410422] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198256, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.415324] env[62368]: DEBUG nova.network.neutron [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.529205] env[62368]: DEBUG nova.scheduler.client.report [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.542895] env[62368]: DEBUG nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.674934] env[62368]: DEBUG nova.network.neutron [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Updating instance_info_cache with network_info: [{"id": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "address": "fa:16:3e:b9:01:76", "network": {"id": "8c5de397-da1e-4836-a5f8-cd85f07a1239", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-831639205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b12e7fa4eec4f3b975651f2fc46c8fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c5d30b5-42", "ovs_interfaceid": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.834219] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198255, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585999} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.834487] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4/9ce5c5a1-5c90-4f76-a511-d8f855b6bab4.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 744.834730] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.835037] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6bd9e0a6-cfb2-4d08-aa68-573c3aef1c58 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.841448] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 744.841448] env[62368]: value = "task-1198257" [ 744.841448] env[62368]: _type = "Task" [ 744.841448] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.849235] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198257, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.877753] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526387e3-f92d-6fc2-0b05-62b59113d101, 'name': SearchDatastore_Task, 'duration_secs': 0.035693} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.882017] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.882017] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b393f654-0611-493e-a89f-445dfd3f8c13/b393f654-0611-493e-a89f-445dfd3f8c13.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 744.882017] env[62368]: DEBUG nova.compute.manager [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Received event network-vif-plugged-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.882017] env[62368]: DEBUG oslo_concurrency.lockutils [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] Acquiring lock "085ebe93-aa24-4626-94fe-241c4297e4db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.882494] env[62368]: DEBUG oslo_concurrency.lockutils [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] Lock "085ebe93-aa24-4626-94fe-241c4297e4db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.882494] env[62368]: DEBUG oslo_concurrency.lockutils [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] Lock "085ebe93-aa24-4626-94fe-241c4297e4db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.882494] env[62368]: DEBUG nova.compute.manager [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] No waiting events found dispatching network-vif-plugged-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 744.882494] env[62368]: WARNING nova.compute.manager [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Received unexpected event network-vif-plugged-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 for instance with vm_state building and task_state spawning. [ 744.882494] env[62368]: DEBUG nova.compute.manager [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Received event network-changed-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.882662] env[62368]: DEBUG nova.compute.manager [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Refreshing instance network info cache due to event network-changed-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.882662] env[62368]: DEBUG oslo_concurrency.lockutils [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] Acquiring lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.882662] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c47a361-dd81-4831-aa60-4e801b53ce1c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.889738] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 744.889738] env[62368]: value = "task-1198258" [ 744.889738] env[62368]: _type = "Task" [ 744.889738] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.899071] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198258, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.909830] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198256, 'name': CreateSnapshot_Task, 'duration_secs': 0.915503} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.911161] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 744.911161] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60b19d9-aa3b-497b-a18d-574d6f0b13d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.033293] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.033717] env[62368]: DEBUG nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.036889] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.038518] env[62368]: INFO nova.compute.claims [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.072911] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.178745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Releasing lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.179106] env[62368]: DEBUG nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Instance network_info: |[{"id": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "address": "fa:16:3e:b9:01:76", "network": {"id": "8c5de397-da1e-4836-a5f8-cd85f07a1239", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-831639205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b12e7fa4eec4f3b975651f2fc46c8fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c5d30b5-42", "ovs_interfaceid": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 745.179439] env[62368]: DEBUG oslo_concurrency.lockutils [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] Acquired lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.179696] env[62368]: DEBUG nova.network.neutron [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Refreshing network info cache for port 3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 745.181072] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:01:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c5d30b5-42b0-4dc4-81a6-b9733f7e0797', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 745.190035] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Creating folder: Project (7b12e7fa4eec4f3b975651f2fc46c8fa). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 745.190517] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-824a16aa-c15e-4038-9fef-140e2ae469a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.206186] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Created folder: Project (7b12e7fa4eec4f3b975651f2fc46c8fa) in parent group-v259706. [ 745.207439] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Creating folder: Instances. Parent ref: group-v259748. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 745.207439] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eeb2368e-e082-4a7c-9c9b-c69b28975b9a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.222695] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Created folder: Instances in parent group-v259748. [ 745.223460] env[62368]: DEBUG oslo.service.loopingcall [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.223460] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 745.223460] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1cf97112-e93d-4d90-92c4-42a00399da94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.246384] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 745.246384] env[62368]: value = "task-1198261" [ 745.246384] env[62368]: _type = "Task" [ 745.246384] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.255846] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198261, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.353180] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198257, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070998} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.353632] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.354367] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5111b1-5c11-4a5f-aef9-e74f96f36857 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.381237] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4/9ce5c5a1-5c90-4f76-a511-d8f855b6bab4.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.381237] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ed20985-c93f-4daa-b10a-0b79b6ceede7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.408149] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198258, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.409818] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 745.409818] env[62368]: value = "task-1198262" [ 745.409818] env[62368]: _type = "Task" [ 745.409818] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.419213] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198262, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.429193] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 745.429595] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-415f5df6-dcd8-4490-855c-b6bd1558f9ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.439703] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 745.439703] env[62368]: value = "task-1198263" [ 745.439703] env[62368]: _type = "Task" [ 745.439703] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.449857] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198263, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.544563] env[62368]: DEBUG nova.compute.utils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.551598] env[62368]: DEBUG nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 745.551598] env[62368]: DEBUG nova.network.neutron [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 745.602677] env[62368]: DEBUG nova.policy [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50e00b1a1fd3401fa803dea5b0801965', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73819709966b4e5f873020b399c49871', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.757367] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198261, 'name': CreateVM_Task, 'duration_secs': 0.390162} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.757567] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 745.758247] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.758408] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.758733] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 745.758984] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b672108-6c6c-4545-95c6-d952c0adea73 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.764063] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 745.764063] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c4f348-0992-4fb1-5607-b738f7c2468c" [ 745.764063] env[62368]: _type = "Task" [ 745.764063] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.778958] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c4f348-0992-4fb1-5607-b738f7c2468c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.908133] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198258, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594941} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.908427] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b393f654-0611-493e-a89f-445dfd3f8c13/b393f654-0611-493e-a89f-445dfd3f8c13.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 745.908641] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 745.908895] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c6245ac-1f10-45a2-b2a0-f87c958b5963 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.923301] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198262, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.924664] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 745.924664] env[62368]: value = "task-1198264" [ 745.924664] env[62368]: _type = "Task" [ 745.924664] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.932359] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198264, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.950413] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198263, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.954068] env[62368]: DEBUG nova.network.neutron [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Updated VIF entry in instance network info cache for port 3c5d30b5-42b0-4dc4-81a6-b9733f7e0797. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 745.954524] env[62368]: DEBUG nova.network.neutron [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Updating instance_info_cache with network_info: [{"id": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "address": "fa:16:3e:b9:01:76", "network": {"id": "8c5de397-da1e-4836-a5f8-cd85f07a1239", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-831639205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b12e7fa4eec4f3b975651f2fc46c8fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c5d30b5-42", "ovs_interfaceid": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.053389] env[62368]: DEBUG nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.064924] env[62368]: DEBUG nova.network.neutron [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Successfully created port: 11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.276734] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c4f348-0992-4fb1-5607-b738f7c2468c, 'name': SearchDatastore_Task, 'duration_secs': 0.032593} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.277678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.278095] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 746.278280] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.278446] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.278629] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 746.279936] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-457e0159-456c-4909-bd8c-3578d210d112 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.295234] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 746.295475] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 746.296394] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c068377-2320-446b-a619-fa9e4ff84bb1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.302323] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 746.302323] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ddae68-019a-30a7-4474-d983fbd2e46e" [ 746.302323] env[62368]: _type = "Task" [ 746.302323] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.318246] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ddae68-019a-30a7-4474-d983fbd2e46e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.404853] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff237a6-8769-4a5e-8c0a-ff108ac84ba0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.412507] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8d1721-3481-42cc-8f02-344d5673d5f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.424661] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198262, 'name': ReconfigVM_Task, 'duration_secs': 0.83768} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.449196] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4/9ce5c5a1-5c90-4f76-a511-d8f855b6bab4.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.450531] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac425126-9f3f-49de-a328-75e57bed9ca6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.459045] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0829153e-f77e-4878-bfb2-ca43a1cd3529 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.463121] env[62368]: DEBUG oslo_concurrency.lockutils [req-583cc6bc-50c5-4912-817b-1fd62c03a712 req-a0457cb4-63c4-4baf-8fdd-ee92a397f4d4 service nova] Releasing lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.471922] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198264, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068066} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.476264] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.476605] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198263, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.476870] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 746.476870] env[62368]: value = "task-1198265" [ 746.476870] env[62368]: _type = "Task" [ 746.476870] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.477610] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65351617-0bb3-4986-bd52-7f5489c97f55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.486475] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04517338-15f8-4cb8-b1ab-d7d50b3e6672 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.508628] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] b393f654-0611-493e-a89f-445dfd3f8c13/b393f654-0611-493e-a89f-445dfd3f8c13.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.512887] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec02b556-f716-4533-808b-abcc3362f774 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.526916] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198265, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.536996] env[62368]: DEBUG nova.compute.provider_tree [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.541384] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 746.541384] env[62368]: value = "task-1198266" [ 746.541384] env[62368]: _type = "Task" [ 746.541384] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.550352] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198266, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.788759] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.789101] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.789383] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.789676] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.789805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.792286] env[62368]: INFO nova.compute.manager [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Terminating instance [ 746.797893] env[62368]: DEBUG nova.compute.manager [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 746.798119] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 746.798938] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c6af93-6e6a-4392-9cf1-bec238c14a79 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.807296] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 746.807553] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cfe9b49-61be-431a-ad6c-aa84c0321259 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.817669] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ddae68-019a-30a7-4474-d983fbd2e46e, 'name': SearchDatastore_Task, 'duration_secs': 0.052212} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.819606] env[62368]: DEBUG oslo_vmware.api [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 746.819606] env[62368]: value = "task-1198267" [ 746.819606] env[62368]: _type = "Task" [ 746.819606] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.819859] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebc2dfd7-c196-47b7-ad36-aedf85d60003 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.827607] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 746.827607] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527d880f-a733-d3f5-db62-7c9117432dd3" [ 746.827607] env[62368]: _type = "Task" [ 746.827607] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.832819] env[62368]: DEBUG oslo_vmware.api [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198267, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.840411] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527d880f-a733-d3f5-db62-7c9117432dd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.960372] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198263, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.991225] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198265, 'name': Rename_Task, 'duration_secs': 0.429109} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.991710] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 746.992154] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e68761e9-8e98-4479-aae3-99cb866bcb5c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.002022] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 747.002022] env[62368]: value = "task-1198268" [ 747.002022] env[62368]: _type = "Task" [ 747.002022] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.008975] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.039627] env[62368]: DEBUG nova.scheduler.client.report [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.061670] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198266, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.067090] env[62368]: DEBUG nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.098930] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.099293] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.099533] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.099778] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.100021] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.100273] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.100571] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.100814] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.101016] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.101263] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.101494] env[62368]: DEBUG nova.virt.hardware [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.106393] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3730ca25-347b-42ec-a3e6-e0dd06b533ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.117201] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763b1d2e-c1e8-4227-935c-91f8efa75cf7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.333223] env[62368]: DEBUG oslo_vmware.api [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198267, 'name': PowerOffVM_Task, 'duration_secs': 0.20383} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.336952] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 747.337265] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 747.337909] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc82e285-530f-47da-a3fa-9da13d628b98 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.345583] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527d880f-a733-d3f5-db62-7c9117432dd3, 'name': SearchDatastore_Task, 'duration_secs': 0.017006} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.345908] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.346232] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 085ebe93-aa24-4626-94fe-241c4297e4db/085ebe93-aa24-4626-94fe-241c4297e4db.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 747.346649] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc96dcea-40bd-4f61-88f0-2193e74353c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.354531] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 747.354531] env[62368]: value = "task-1198270" [ 747.354531] env[62368]: _type = "Task" [ 747.354531] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.362873] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.417317] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 747.417772] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 747.418062] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Deleting the datastore file [datastore1] 4dc7b6b4-1bf5-4195-bb93-14756f8f9986 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.418476] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3625e39-273c-4dac-bdb9-647ac0a293c9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.429858] env[62368]: DEBUG oslo_vmware.api [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for the task: (returnval){ [ 747.429858] env[62368]: value = "task-1198271" [ 747.429858] env[62368]: _type = "Task" [ 747.429858] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.439947] env[62368]: DEBUG oslo_vmware.api [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.460803] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198263, 'name': CloneVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.510874] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198268, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.552846] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.553901] env[62368]: DEBUG nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.563902] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.815s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.565779] env[62368]: INFO nova.compute.claims [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.569858] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198266, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.715316] env[62368]: DEBUG nova.compute.manager [req-311d1ce9-94bd-4f4b-8af8-967352a1c1d1 req-6da846bb-8e89-4e5b-952c-af14ac37c243 service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Received event network-vif-plugged-11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.715643] env[62368]: DEBUG oslo_concurrency.lockutils [req-311d1ce9-94bd-4f4b-8af8-967352a1c1d1 req-6da846bb-8e89-4e5b-952c-af14ac37c243 service nova] Acquiring lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.715879] env[62368]: DEBUG oslo_concurrency.lockutils [req-311d1ce9-94bd-4f4b-8af8-967352a1c1d1 req-6da846bb-8e89-4e5b-952c-af14ac37c243 service nova] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.716313] env[62368]: DEBUG oslo_concurrency.lockutils [req-311d1ce9-94bd-4f4b-8af8-967352a1c1d1 req-6da846bb-8e89-4e5b-952c-af14ac37c243 service nova] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.716457] env[62368]: DEBUG nova.compute.manager [req-311d1ce9-94bd-4f4b-8af8-967352a1c1d1 req-6da846bb-8e89-4e5b-952c-af14ac37c243 service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] No waiting events found dispatching network-vif-plugged-11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 747.716663] env[62368]: WARNING nova.compute.manager [req-311d1ce9-94bd-4f4b-8af8-967352a1c1d1 req-6da846bb-8e89-4e5b-952c-af14ac37c243 service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Received unexpected event network-vif-plugged-11ce4c60-0df0-4c43-bbe4-6cc53717309b for instance with vm_state building and task_state spawning. [ 747.800479] env[62368]: DEBUG nova.network.neutron [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Successfully updated port: 11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 747.866870] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198270, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.941490] env[62368]: DEBUG oslo_vmware.api [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Task: {'id': task-1198271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.499398} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.944783] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 747.944783] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 747.944783] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 747.944783] env[62368]: INFO nova.compute.manager [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Took 1.14 seconds to destroy the instance on the hypervisor. [ 747.944783] env[62368]: DEBUG oslo.service.loopingcall [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.945361] env[62368]: DEBUG nova.compute.manager [-] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.945361] env[62368]: DEBUG nova.network.neutron [-] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 747.961224] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198263, 'name': CloneVM_Task, 'duration_secs': 2.08234} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.961921] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Created linked-clone VM from snapshot [ 747.962837] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77299ca2-dc26-444e-96c1-3984f6c33584 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.971729] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Uploading image 493bfe5d-9953-4198-b6b8-c770ccec8c9f {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 747.986369] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 747.986369] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-51b509d5-b585-47f0-805b-80b5bd2a803e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.993959] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 747.993959] env[62368]: value = "task-1198272" [ 747.993959] env[62368]: _type = "Task" [ 747.993959] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.008436] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198272, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.013319] env[62368]: DEBUG oslo_vmware.api [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198268, 'name': PowerOnVM_Task, 'duration_secs': 0.57624} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.013848] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 748.014240] env[62368]: INFO nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Took 10.40 seconds to spawn the instance on the hypervisor. [ 748.014680] env[62368]: DEBUG nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 748.015508] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df0063f-725e-459b-9b1b-6421109c2368 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.055625] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198266, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.060195] env[62368]: DEBUG nova.compute.utils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.061590] env[62368]: DEBUG nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.061750] env[62368]: DEBUG nova.network.neutron [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 748.118169] env[62368]: DEBUG nova.policy [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad1a9bdf4b91418ea60a18d4e2f34ef3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b87e3b7d1d947b79f4834e4e0fe0395', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.309598] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.309598] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.309598] env[62368]: DEBUG nova.network.neutron [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.370421] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198270, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551375} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.370908] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 085ebe93-aa24-4626-94fe-241c4297e4db/085ebe93-aa24-4626-94fe-241c4297e4db.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 748.371547] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 748.372107] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88e19e56-35d3-4ebf-b68b-6124d587a208 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.380234] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 748.380234] env[62368]: value = "task-1198273" [ 748.380234] env[62368]: _type = "Task" [ 748.380234] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.391603] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198273, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.447700] env[62368]: DEBUG nova.network.neutron [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Successfully created port: 280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.506910] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198272, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.537140] env[62368]: INFO nova.compute.manager [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Took 30.76 seconds to build instance. [ 748.557431] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198266, 'name': ReconfigVM_Task, 'duration_secs': 1.609148} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.558328] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Reconfigured VM instance instance-00000030 to attach disk [datastore1] b393f654-0611-493e-a89f-445dfd3f8c13/b393f654-0611-493e-a89f-445dfd3f8c13.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.558982] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-995163d0-7742-4772-8862-d42911a70251 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.566013] env[62368]: DEBUG nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.570409] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 748.570409] env[62368]: value = "task-1198274" [ 748.570409] env[62368]: _type = "Task" [ 748.570409] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.591149] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198274, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.863336] env[62368]: DEBUG nova.network.neutron [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.893671] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198273, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069587} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.893991] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 748.894865] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f140ee-e968-4a0f-a550-fd9f06dc4b96 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.899631] env[62368]: DEBUG nova.network.neutron [-] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.924534] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 085ebe93-aa24-4626-94fe-241c4297e4db/085ebe93-aa24-4626-94fe-241c4297e4db.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 748.928214] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1b86c12-66e9-4291-872a-cf312e21e870 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.952276] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 748.952276] env[62368]: value = "task-1198275" [ 748.952276] env[62368]: _type = "Task" [ 748.952276] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.961964] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198275, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.967142] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e293e8-8cd9-4ce5-a197-021d4651503c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.972192] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1d5681-1529-432d-b90f-eabb00241e35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.009462] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7511ab1a-bad3-4bca-86d7-08eeca124792 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.021549] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198272, 'name': Destroy_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.022892] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522ce8b1-7ae1-4531-9722-303400590d5d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.038854] env[62368]: DEBUG nova.compute.provider_tree [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.040206] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a2954933-60f3-4974-87e0-7a6a82ea7b7b tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.356s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.097572] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198274, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.202781] env[62368]: DEBUG nova.network.neutron [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.402769] env[62368]: INFO nova.compute.manager [-] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Took 1.46 seconds to deallocate network for instance. [ 749.464903] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198275, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.515505] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198272, 'name': Destroy_Task, 'duration_secs': 1.127871} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.515859] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Destroyed the VM [ 749.516293] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 749.516550] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6eeaa2ed-cfce-4898-93ac-ca1494ce5c05 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.522667] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 749.522667] env[62368]: value = "task-1198276" [ 749.522667] env[62368]: _type = "Task" [ 749.522667] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.530329] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198276, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.545019] env[62368]: DEBUG nova.scheduler.client.report [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.546050] env[62368]: DEBUG nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.587592] env[62368]: DEBUG nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.593728] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198274, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.610820] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.611085] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.611248] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.611462] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.611612] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.611759] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.611962] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.612135] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.612334] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.612460] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.612633] env[62368]: DEBUG nova.virt.hardware [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.613537] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b21edd9-812a-49c0-b44e-4dea8a1972ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.620963] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf61d5e-2204-48c0-8495-564c73efe030 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.704940] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.705383] env[62368]: DEBUG nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Instance network_info: |[{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 749.705842] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:9e:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0685bd0b-3dbf-4a06-951c-c6a4726dd4b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '11ce4c60-0df0-4c43-bbe4-6cc53717309b', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.719877] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating folder: Project (73819709966b4e5f873020b399c49871). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 749.720075] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aee317ec-6b07-4e3d-8cad-63b8af84138e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.734064] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Created folder: Project (73819709966b4e5f873020b399c49871) in parent group-v259706. [ 749.734283] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating folder: Instances. Parent ref: group-v259752. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 749.734519] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f6c91579-5c4d-4bbe-a711-58ec49652ae1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.745000] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Created folder: Instances in parent group-v259752. [ 749.745331] env[62368]: DEBUG oslo.service.loopingcall [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.745503] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 749.745714] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2c714fa-0c12-42b8-82fc-d911a49e61b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.767265] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.767265] env[62368]: value = "task-1198279" [ 749.767265] env[62368]: _type = "Task" [ 749.767265] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.775383] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198279, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.856961] env[62368]: DEBUG nova.compute.manager [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Received event network-changed-11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.860100] env[62368]: DEBUG nova.compute.manager [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Refreshing instance network info cache due to event network-changed-11ce4c60-0df0-4c43-bbe4-6cc53717309b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 749.860100] env[62368]: DEBUG oslo_concurrency.lockutils [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] Acquiring lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.860100] env[62368]: DEBUG oslo_concurrency.lockutils [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] Acquired lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.860100] env[62368]: DEBUG nova.network.neutron [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Refreshing network info cache for port 11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 749.909689] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.963449] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198275, 'name': ReconfigVM_Task, 'duration_secs': 0.802493} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.963880] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 085ebe93-aa24-4626-94fe-241c4297e4db/085ebe93-aa24-4626-94fe-241c4297e4db.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 749.964672] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d4b6639-58d8-448a-9981-8dd2143680a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.973122] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 749.973122] env[62368]: value = "task-1198280" [ 749.973122] env[62368]: _type = "Task" [ 749.973122] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.981103] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198280, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.035688] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198276, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.049889] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.053016] env[62368]: DEBUG nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.059412] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.759s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.059956] env[62368]: DEBUG nova.objects.instance [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lazy-loading 'resources' on Instance uuid a5c638d8-f761-497d-bde2-39ea90d8d151 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.083336] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.097640] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198274, 'name': Rename_Task, 'duration_secs': 1.280489} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.097640] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 750.097640] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f295d2d5-a889-4330-af3f-1c5b7cbee317 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.103054] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 750.103054] env[62368]: value = "task-1198281" [ 750.103054] env[62368]: _type = "Task" [ 750.103054] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.113964] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198281, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.187926] env[62368]: DEBUG nova.network.neutron [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Successfully updated port: 280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 750.191510] env[62368]: DEBUG nova.compute.manager [req-fe458f90-5863-4b90-b616-aee9be06ad60 req-05fbdef9-3fce-401d-a92b-8fa0a42d1c3a service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Received event network-vif-plugged-280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.191510] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe458f90-5863-4b90-b616-aee9be06ad60 req-05fbdef9-3fce-401d-a92b-8fa0a42d1c3a service nova] Acquiring lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.191510] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe458f90-5863-4b90-b616-aee9be06ad60 req-05fbdef9-3fce-401d-a92b-8fa0a42d1c3a service nova] Lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.191510] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe458f90-5863-4b90-b616-aee9be06ad60 req-05fbdef9-3fce-401d-a92b-8fa0a42d1c3a service nova] Lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.191510] env[62368]: DEBUG nova.compute.manager [req-fe458f90-5863-4b90-b616-aee9be06ad60 req-05fbdef9-3fce-401d-a92b-8fa0a42d1c3a service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] No waiting events found dispatching network-vif-plugged-280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 750.192026] env[62368]: WARNING nova.compute.manager [req-fe458f90-5863-4b90-b616-aee9be06ad60 req-05fbdef9-3fce-401d-a92b-8fa0a42d1c3a service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Received unexpected event network-vif-plugged-280f8ef3-0f46-4846-9e44-fff59d473e8b for instance with vm_state building and task_state spawning. [ 750.285462] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198279, 'name': CreateVM_Task, 'duration_secs': 0.387032} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.285462] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 750.285462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.285462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.285462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 750.285462] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5c2f488-4561-4e23-87fe-fd082af581d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.289909] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 750.289909] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52b7abfd-a3a4-f32b-140e-86beaf8e0899" [ 750.289909] env[62368]: _type = "Task" [ 750.289909] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.301917] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b7abfd-a3a4-f32b-140e-86beaf8e0899, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.484808] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198280, 'name': Rename_Task, 'duration_secs': 0.148619} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.485123] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 750.485546] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d0ba78e-35e5-4592-a593-27a77a998c53 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.491781] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 750.491781] env[62368]: value = "task-1198282" [ 750.491781] env[62368]: _type = "Task" [ 750.491781] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.500126] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198282, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.507093] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "eea21546-fbbf-4440-829c-8583c4ccabb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.507093] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.534685] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198276, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.561330] env[62368]: DEBUG nova.compute.utils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.563345] env[62368]: DEBUG nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 750.563746] env[62368]: DEBUG nova.network.neutron [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 750.619117] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198281, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.635041] env[62368]: DEBUG nova.policy [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f3afe492e7e47f28c3c8dad61f819d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81b69987192149ddb6dfcafc1f7770fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 750.644882] env[62368]: DEBUG nova.network.neutron [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updated VIF entry in instance network info cache for port 11ce4c60-0df0-4c43-bbe4-6cc53717309b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 750.645297] env[62368]: DEBUG nova.network.neutron [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.693134] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.693134] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquired lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.693134] env[62368]: DEBUG nova.network.neutron [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.807458] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b7abfd-a3a4-f32b-140e-86beaf8e0899, 'name': SearchDatastore_Task, 'duration_secs': 0.010978} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.807458] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.807458] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 750.807458] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.807800] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.807800] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 750.807800] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15d5210c-a4ec-4060-b83f-483abfc700e0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.820593] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 750.820593] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 750.821376] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac1c92f9-816f-4933-b733-eac5c7db93bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.826835] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 750.826835] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526a64a0-5e1f-3c7b-480b-9b5db86bfdf2" [ 750.826835] env[62368]: _type = "Task" [ 750.826835] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.835644] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526a64a0-5e1f-3c7b-480b-9b5db86bfdf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.978432] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4a18d6-3342-4a17-a820-97deb936ad01 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.986452] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9808b2e8-59e7-4ef0-ac16-8a193f739602 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.019629] env[62368]: DEBUG nova.network.neutron [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Successfully created port: 6e4af274-eb1c-4a81-b1f6-dc415b7be3d5 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.024840] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d114977b-5fc5-41c5-b11f-48ff328834e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.035179] env[62368]: DEBUG oslo_vmware.api [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198282, 'name': PowerOnVM_Task, 'duration_secs': 0.489877} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.038971] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 751.038971] env[62368]: INFO nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Took 8.36 seconds to spawn the instance on the hypervisor. [ 751.038971] env[62368]: DEBUG nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 751.038971] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d2ed8b-d5c0-4fff-b008-8ccc556f6f9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.041613] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a14c5b-13a3-4264-9c5a-191714e1958d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.049251] env[62368]: DEBUG oslo_vmware.api [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198276, 'name': RemoveSnapshot_Task, 'duration_secs': 1.365344} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.049251] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 751.061760] env[62368]: DEBUG nova.compute.provider_tree [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.069024] env[62368]: DEBUG nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.117230] env[62368]: DEBUG oslo_vmware.api [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198281, 'name': PowerOnVM_Task, 'duration_secs': 0.622456} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.117430] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 751.117592] env[62368]: INFO nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Took 10.91 seconds to spawn the instance on the hypervisor. [ 751.117767] env[62368]: DEBUG nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 751.118862] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c66186-ff71-4a42-a220-df3ba0786207 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.152023] env[62368]: DEBUG oslo_concurrency.lockutils [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] Releasing lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.152023] env[62368]: DEBUG nova.compute.manager [req-d4640ed6-f9ea-4f98-a06d-ba5231ef2d1e req-e75e61d2-7102-43f7-8383-46dda1922f7b service nova] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Received event network-vif-deleted-fe5fd04f-da09-4430-91aa-a990e3a3f4df {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.233971] env[62368]: DEBUG nova.network.neutron [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.337559] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526a64a0-5e1f-3c7b-480b-9b5db86bfdf2, 'name': SearchDatastore_Task, 'duration_secs': 0.011552} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.338370] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f7b452d-1934-4235-a302-605296e0f6bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.343595] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 751.343595] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e83b07-fe99-8460-7e88-d2536a010c97" [ 751.343595] env[62368]: _type = "Task" [ 751.343595] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.353254] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e83b07-fe99-8460-7e88-d2536a010c97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.393306] env[62368]: DEBUG nova.network.neutron [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updating instance_info_cache with network_info: [{"id": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "address": "fa:16:3e:fc:46:9c", "network": {"id": "d71a42e9-1e1c-4abd-b585-2362ba1b8cfb", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1629701000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3b87e3b7d1d947b79f4834e4e0fe0395", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280f8ef3-0f", "ovs_interfaceid": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.564329] env[62368]: DEBUG nova.scheduler.client.report [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.568929] env[62368]: WARNING nova.compute.manager [None req-a02a06d8-0583-4078-b5d5-b714703d54e0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Image not found during snapshot: nova.exception.ImageNotFound: Image 493bfe5d-9953-4198-b6b8-c770ccec8c9f could not be found. [ 751.588498] env[62368]: INFO nova.compute.manager [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Took 31.09 seconds to build instance. [ 751.639516] env[62368]: INFO nova.compute.manager [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Took 32.64 seconds to build instance. [ 751.855471] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e83b07-fe99-8460-7e88-d2536a010c97, 'name': SearchDatastore_Task, 'duration_secs': 0.066258} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.855744] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.856709] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 13765305-2e55-4ee8-9a6f-4ae5ee724367/13765305-2e55-4ee8-9a6f-4ae5ee724367.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 751.857229] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12da7fb3-9ec7-4ba4-a4c0-af32987f5f8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.866023] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 751.866023] env[62368]: value = "task-1198283" [ 751.866023] env[62368]: _type = "Task" [ 751.866023] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.878381] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198283, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.896048] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Releasing lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.896369] env[62368]: DEBUG nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Instance network_info: |[{"id": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "address": "fa:16:3e:fc:46:9c", "network": {"id": "d71a42e9-1e1c-4abd-b585-2362ba1b8cfb", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1629701000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3b87e3b7d1d947b79f4834e4e0fe0395", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280f8ef3-0f", "ovs_interfaceid": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 751.896771] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:46:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f8442aa5-73db-4599-8564-b98a6ea26b9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '280f8ef3-0f46-4846-9e44-fff59d473e8b', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.906342] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Creating folder: Project (3b87e3b7d1d947b79f4834e4e0fe0395). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 751.906667] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4cdd3259-9423-4301-b9ca-88eabb7f7706 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.918660] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Created folder: Project (3b87e3b7d1d947b79f4834e4e0fe0395) in parent group-v259706. [ 751.918868] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Creating folder: Instances. Parent ref: group-v259755. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 751.919135] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ec7e4e3c-e94f-446e-8a38-0719e93f91a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.928530] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Created folder: Instances in parent group-v259755. [ 751.928915] env[62368]: DEBUG oslo.service.loopingcall [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.929198] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 751.929437] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9368057b-3209-49ca-9df8-ad6f77839350 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.955185] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.955185] env[62368]: value = "task-1198286" [ 751.955185] env[62368]: _type = "Task" [ 751.955185] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.962859] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198286, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.082582] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.023s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.085498] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.576s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.087454] env[62368]: INFO nova.compute.claims [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.090979] env[62368]: DEBUG nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.093676] env[62368]: DEBUG oslo_concurrency.lockutils [None req-127cbd80-bbf3-4ace-aa50-2ab9ba882b29 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "085ebe93-aa24-4626-94fe-241c4297e4db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.492s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.093991] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.094234] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.094603] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.094675] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.094832] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.098501] env[62368]: INFO nova.compute.manager [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Terminating instance [ 752.100437] env[62368]: DEBUG nova.compute.manager [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.100536] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 752.101347] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f413d64-5a70-4d64-a14b-14ae5897147a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.110097] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 752.111929] env[62368]: INFO nova.scheduler.client.report [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Deleted allocations for instance a5c638d8-f761-497d-bde2-39ea90d8d151 [ 752.112901] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8614ac7f-e465-46ae-8a92-bf00f5a59d2c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.127859] env[62368]: DEBUG oslo_vmware.api [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 752.127859] env[62368]: value = "task-1198287" [ 752.127859] env[62368]: _type = "Task" [ 752.127859] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.130863] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.131118] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.131282] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.131484] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.131628] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.131776] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.131988] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.132161] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.132329] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.132492] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.132659] env[62368]: DEBUG nova.virt.hardware [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.134085] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381fccda-3f70-433e-b193-e5643072169c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.141048] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8eedd2a6-0544-4e95-b0f4-a5f287c01f81 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "b393f654-0611-493e-a89f-445dfd3f8c13" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.784s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.150128] env[62368]: DEBUG oslo_vmware.api [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.152212] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4ca670-2b6f-41b5-8b89-bbcb5370cd96 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.380553] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198283, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.418711] env[62368]: DEBUG nova.compute.manager [req-c0497417-5ac1-4dab-82e3-e6016c04cccf req-e4777e7c-bb22-4796-ac3f-20ded6ee53e7 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Received event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.418711] env[62368]: DEBUG nova.compute.manager [req-c0497417-5ac1-4dab-82e3-e6016c04cccf req-e4777e7c-bb22-4796-ac3f-20ded6ee53e7 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing instance network info cache due to event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 752.419152] env[62368]: DEBUG oslo_concurrency.lockutils [req-c0497417-5ac1-4dab-82e3-e6016c04cccf req-e4777e7c-bb22-4796-ac3f-20ded6ee53e7 service nova] Acquiring lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.419377] env[62368]: DEBUG oslo_concurrency.lockutils [req-c0497417-5ac1-4dab-82e3-e6016c04cccf req-e4777e7c-bb22-4796-ac3f-20ded6ee53e7 service nova] Acquired lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.419560] env[62368]: DEBUG nova.network.neutron [req-c0497417-5ac1-4dab-82e3-e6016c04cccf req-e4777e7c-bb22-4796-ac3f-20ded6ee53e7 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 752.466054] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198286, 'name': CreateVM_Task, 'duration_secs': 0.353331} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.466240] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 752.466973] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.467460] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.467797] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.468072] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6a22e63-2b32-4501-9439-669716d1aeaa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.472783] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 752.472783] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c0ebc4-99fc-e770-9bae-fb3c52cf23c2" [ 752.472783] env[62368]: _type = "Task" [ 752.472783] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.480714] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c0ebc4-99fc-e770-9bae-fb3c52cf23c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.600341] env[62368]: DEBUG nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.626110] env[62368]: DEBUG oslo_concurrency.lockutils [None req-80115cd4-2844-48e7-9da1-ca84ca49ed97 tempest-ServersAaction247Test-176263630 tempest-ServersAaction247Test-176263630-project-member] Lock "a5c638d8-f761-497d-bde2-39ea90d8d151" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.208s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.643504] env[62368]: DEBUG oslo_vmware.api [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198287, 'name': PowerOffVM_Task, 'duration_secs': 0.296212} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.643725] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 752.643888] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 752.644146] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc3e1aaa-f63a-440c-abf8-e214e3f593fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.652795] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.715022] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 752.715022] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 752.715022] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleting the datastore file [datastore1] e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 752.715022] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f0b44de-e3fd-42d0-b0da-70c03e35996c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.722733] env[62368]: DEBUG oslo_vmware.api [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 752.722733] env[62368]: value = "task-1198289" [ 752.722733] env[62368]: _type = "Task" [ 752.722733] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.730815] env[62368]: DEBUG oslo_vmware.api [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.880111] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198283, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.600565} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.880369] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 13765305-2e55-4ee8-9a6f-4ae5ee724367/13765305-2e55-4ee8-9a6f-4ae5ee724367.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 752.880630] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 752.880892] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00db8aa6-2608-4da1-8dc2-6fa7d15b0784 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.886264] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 752.886264] env[62368]: value = "task-1198290" [ 752.886264] env[62368]: _type = "Task" [ 752.886264] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.898409] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198290, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.941112] env[62368]: DEBUG nova.network.neutron [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Successfully updated port: 6e4af274-eb1c-4a81-b1f6-dc415b7be3d5 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 752.984940] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c0ebc4-99fc-e770-9bae-fb3c52cf23c2, 'name': SearchDatastore_Task, 'duration_secs': 0.008152} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.985256] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.985486] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.985714] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.985851] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.986027] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.986279] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f72e5a2-c530-49b5-bc12-97df21df52a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.994722] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.994938] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 752.995690] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b4e342a-05e6-4b18-80e6-4d90dd496895 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.000842] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 753.000842] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]528e9ffd-b8c1-fc93-00d6-ed5870dc00f6" [ 753.000842] env[62368]: _type = "Task" [ 753.000842] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.008787] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528e9ffd-b8c1-fc93-00d6-ed5870dc00f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.127163] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.161032] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquiring lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.161032] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.161032] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquiring lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.161361] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.161675] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.165628] env[62368]: INFO nova.compute.manager [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Terminating instance [ 753.169625] env[62368]: DEBUG nova.compute.manager [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 753.169942] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 753.171537] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e03f3c-941e-49be-a20e-3f154769a70f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.180125] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 753.182833] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.183194] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c72986c2-38e6-49e4-b7a6-3bf6a9274045 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.191021] env[62368]: DEBUG oslo_vmware.api [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 753.191021] env[62368]: value = "task-1198291" [ 753.191021] env[62368]: _type = "Task" [ 753.191021] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.197623] env[62368]: DEBUG oslo_vmware.api [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198291, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.202278] env[62368]: DEBUG nova.network.neutron [req-c0497417-5ac1-4dab-82e3-e6016c04cccf req-e4777e7c-bb22-4796-ac3f-20ded6ee53e7 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updated VIF entry in instance network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 753.202769] env[62368]: DEBUG nova.network.neutron [req-c0497417-5ac1-4dab-82e3-e6016c04cccf req-e4777e7c-bb22-4796-ac3f-20ded6ee53e7 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updating instance_info_cache with network_info: [{"id": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "address": "fa:16:3e:fc:46:9c", "network": {"id": "d71a42e9-1e1c-4abd-b585-2362ba1b8cfb", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1629701000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3b87e3b7d1d947b79f4834e4e0fe0395", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280f8ef3-0f", "ovs_interfaceid": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.243020] env[62368]: DEBUG oslo_vmware.api [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198289, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157591} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.243383] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 753.243912] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 753.244361] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 753.244758] env[62368]: INFO nova.compute.manager [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 753.245151] env[62368]: DEBUG oslo.service.loopingcall [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.245926] env[62368]: DEBUG nova.compute.manager [-] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.246162] env[62368]: DEBUG nova.network.neutron [-] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.265384] env[62368]: DEBUG nova.compute.manager [req-ec221413-c19d-4870-bfce-2cca78bbdd7e req-9c3fbce4-09c4-4da9-8993-e6dc07e0ebae service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Received event network-vif-plugged-6e4af274-eb1c-4a81-b1f6-dc415b7be3d5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.265711] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec221413-c19d-4870-bfce-2cca78bbdd7e req-9c3fbce4-09c4-4da9-8993-e6dc07e0ebae service nova] Acquiring lock "417f000f-cf23-404d-877c-45990d1a7c77-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.266084] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec221413-c19d-4870-bfce-2cca78bbdd7e req-9c3fbce4-09c4-4da9-8993-e6dc07e0ebae service nova] Lock "417f000f-cf23-404d-877c-45990d1a7c77-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.266183] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec221413-c19d-4870-bfce-2cca78bbdd7e req-9c3fbce4-09c4-4da9-8993-e6dc07e0ebae service nova] Lock "417f000f-cf23-404d-877c-45990d1a7c77-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.266363] env[62368]: DEBUG nova.compute.manager [req-ec221413-c19d-4870-bfce-2cca78bbdd7e req-9c3fbce4-09c4-4da9-8993-e6dc07e0ebae service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] No waiting events found dispatching network-vif-plugged-6e4af274-eb1c-4a81-b1f6-dc415b7be3d5 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 753.266542] env[62368]: WARNING nova.compute.manager [req-ec221413-c19d-4870-bfce-2cca78bbdd7e req-9c3fbce4-09c4-4da9-8993-e6dc07e0ebae service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Received unexpected event network-vif-plugged-6e4af274-eb1c-4a81-b1f6-dc415b7be3d5 for instance with vm_state building and task_state spawning. [ 753.270718] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquiring lock "b393f654-0611-493e-a89f-445dfd3f8c13" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.270947] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "b393f654-0611-493e-a89f-445dfd3f8c13" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.271676] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquiring lock "b393f654-0611-493e-a89f-445dfd3f8c13-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.271895] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "b393f654-0611-493e-a89f-445dfd3f8c13-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.272113] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "b393f654-0611-493e-a89f-445dfd3f8c13-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.274281] env[62368]: INFO nova.compute.manager [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Terminating instance [ 753.280702] env[62368]: DEBUG nova.compute.manager [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 753.280702] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 753.281928] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5872e206-0d14-4e7c-85a9-0ffb7a5cbd09 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.290312] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 753.290312] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-988e82fc-3124-472c-bce3-73740df99bc9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.298864] env[62368]: DEBUG oslo_vmware.api [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 753.298864] env[62368]: value = "task-1198292" [ 753.298864] env[62368]: _type = "Task" [ 753.298864] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.308909] env[62368]: DEBUG oslo_vmware.api [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.399509] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198290, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063455} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.399509] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 753.400143] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5600cf88-17e4-4d12-8412-e24400bcd98f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.430617] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 13765305-2e55-4ee8-9a6f-4ae5ee724367/13765305-2e55-4ee8-9a6f-4ae5ee724367.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 753.433675] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43ff0e4a-f5ef-4426-a50c-85f18b30cab6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.453606] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "refresh_cache-417f000f-cf23-404d-877c-45990d1a7c77" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.453606] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "refresh_cache-417f000f-cf23-404d-877c-45990d1a7c77" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.453606] env[62368]: DEBUG nova.network.neutron [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.457190] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 753.457190] env[62368]: value = "task-1198293" [ 753.457190] env[62368]: _type = "Task" [ 753.457190] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.470885] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198293, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.524144] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528e9ffd-b8c1-fc93-00d6-ed5870dc00f6, 'name': SearchDatastore_Task, 'duration_secs': 0.008535} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.525147] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d74e0cf7-76f4-4e7b-b475-18f558e3d4dd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.534702] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb186c5-4589-4e19-b453-786e8bbe2869 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.540165] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 753.540165] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52666935-55b9-c7ba-2943-01d20c5bd877" [ 753.540165] env[62368]: _type = "Task" [ 753.540165] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.548777] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b2d6f1-be35-40b2-aa1b-7560b33d1726 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.559022] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52666935-55b9-c7ba-2943-01d20c5bd877, 'name': SearchDatastore_Task, 'duration_secs': 0.014143} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.559940] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.560767] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d/4fe7b1a2-f894-4131-9456-fb4df8a4532d.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 753.560767] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b77170c1-477f-4864-a397-35a02a9d2611 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.591021] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5a1d61-a943-4abb-897d-72ca51449980 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.595703] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 753.595703] env[62368]: value = "task-1198294" [ 753.595703] env[62368]: _type = "Task" [ 753.595703] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.603807] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb7fdec-66af-4613-a0ba-a8564634fa86 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.611186] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198294, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.621341] env[62368]: DEBUG nova.compute.provider_tree [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.701332] env[62368]: DEBUG oslo_vmware.api [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198291, 'name': PowerOffVM_Task, 'duration_secs': 0.400978} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.701624] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 753.701858] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 753.702333] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ac92c16-3b0d-4602-86ed-e67912189980 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.705737] env[62368]: DEBUG oslo_concurrency.lockutils [req-c0497417-5ac1-4dab-82e3-e6016c04cccf req-e4777e7c-bb22-4796-ac3f-20ded6ee53e7 service nova] Releasing lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.771682] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 753.772186] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 753.772679] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Deleting the datastore file [datastore1] 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 753.772679] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17591b62-2e0d-4afa-8409-8cf4c48517df {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.779981] env[62368]: DEBUG oslo_vmware.api [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for the task: (returnval){ [ 753.779981] env[62368]: value = "task-1198296" [ 753.779981] env[62368]: _type = "Task" [ 753.779981] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.789988] env[62368]: DEBUG oslo_vmware.api [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198296, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.809625] env[62368]: DEBUG oslo_vmware.api [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198292, 'name': PowerOffVM_Task, 'duration_secs': 0.347153} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.809994] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 753.810306] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 753.810645] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56108036-8e43-4570-9768-34ed2466894b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.937567] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 753.937881] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 753.938219] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Deleting the datastore file [datastore1] b393f654-0611-493e-a89f-445dfd3f8c13 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 753.938508] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e579cf0-e24d-4335-8f92-64d5eefa0f2c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.945691] env[62368]: DEBUG oslo_vmware.api [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for the task: (returnval){ [ 753.945691] env[62368]: value = "task-1198298" [ 753.945691] env[62368]: _type = "Task" [ 753.945691] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.955054] env[62368]: DEBUG oslo_vmware.api [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198298, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.971135] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198293, 'name': ReconfigVM_Task, 'duration_secs': 0.394903} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.971476] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 13765305-2e55-4ee8-9a6f-4ae5ee724367/13765305-2e55-4ee8-9a6f-4ae5ee724367.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 753.975025] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb553568-dd35-4bdc-9fa4-60696c1f42f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.979283] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 753.979283] env[62368]: value = "task-1198299" [ 753.979283] env[62368]: _type = "Task" [ 753.979283] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.989620] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198299, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.009367] env[62368]: DEBUG nova.network.neutron [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.061625] env[62368]: DEBUG nova.network.neutron [-] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.116378] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198294, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.125127] env[62368]: DEBUG nova.scheduler.client.report [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.291590] env[62368]: DEBUG oslo_vmware.api [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Task: {'id': task-1198296, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.327826} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.291832] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.292025] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 754.292951] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 754.292951] env[62368]: INFO nova.compute.manager [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Took 1.12 seconds to destroy the instance on the hypervisor. [ 754.292951] env[62368]: DEBUG oslo.service.loopingcall [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.293217] env[62368]: DEBUG nova.compute.manager [-] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.293217] env[62368]: DEBUG nova.network.neutron [-] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 754.326503] env[62368]: DEBUG nova.network.neutron [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Updating instance_info_cache with network_info: [{"id": "6e4af274-eb1c-4a81-b1f6-dc415b7be3d5", "address": "fa:16:3e:2d:d5:37", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e4af274-eb", "ovs_interfaceid": "6e4af274-eb1c-4a81-b1f6-dc415b7be3d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.457764] env[62368]: DEBUG oslo_vmware.api [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198298, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.489576] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198299, 'name': Rename_Task, 'duration_secs': 0.222136} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.489954] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 754.490732] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-954ea70f-61d9-4c03-968b-8ab3ebf06c5b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.498037] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 754.498037] env[62368]: value = "task-1198300" [ 754.498037] env[62368]: _type = "Task" [ 754.498037] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.507047] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198300, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.566787] env[62368]: INFO nova.compute.manager [-] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Took 1.32 seconds to deallocate network for instance. [ 754.607564] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198294, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60908} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.608407] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d/4fe7b1a2-f894-4131-9456-fb4df8a4532d.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 754.608407] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 754.609999] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc1cea8b-beae-4229-881b-54f9a7bc161b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.620255] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 754.620255] env[62368]: value = "task-1198301" [ 754.620255] env[62368]: _type = "Task" [ 754.620255] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.629045] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198301, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.633357] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.634568] env[62368]: DEBUG nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.637063] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.377s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.830304] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "refresh_cache-417f000f-cf23-404d-877c-45990d1a7c77" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.830573] env[62368]: DEBUG nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Instance network_info: |[{"id": "6e4af274-eb1c-4a81-b1f6-dc415b7be3d5", "address": "fa:16:3e:2d:d5:37", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e4af274-eb", "ovs_interfaceid": "6e4af274-eb1c-4a81-b1f6-dc415b7be3d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 754.830999] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:d5:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e4af274-eb1c-4a81-b1f6-dc415b7be3d5', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.838720] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating folder: Project (81b69987192149ddb6dfcafc1f7770fb). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 754.840086] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a031b00-cb06-4b2e-a85c-fb0a2a0a4ab9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.843111] env[62368]: DEBUG nova.compute.manager [req-4300c8c3-de30-45b2-a798-5a00c56ab3c8 req-0cc4712b-a33a-4fe6-869c-bca6befbd404 service nova] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Received event network-vif-deleted-4ca48f55-7ae9-49a3-917d-cb0a6eec5043 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.855258] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Created folder: Project (81b69987192149ddb6dfcafc1f7770fb) in parent group-v259706. [ 754.855500] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating folder: Instances. Parent ref: group-v259758. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 754.855773] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cfaa3b9c-a4d5-452f-889e-655e1c5ad25b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.866110] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Created folder: Instances in parent group-v259758. [ 754.866401] env[62368]: DEBUG oslo.service.loopingcall [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.866610] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 754.866815] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20bef5c1-aac9-4bc7-8133-bfbd2459efa9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.887567] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.887567] env[62368]: value = "task-1198304" [ 754.887567] env[62368]: _type = "Task" [ 754.887567] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.895939] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198304, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.958223] env[62368]: DEBUG oslo_vmware.api [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Task: {'id': task-1198298, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.556469} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.958223] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.958223] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 754.958223] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 754.958443] env[62368]: INFO nova.compute.manager [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Took 1.68 seconds to destroy the instance on the hypervisor. [ 754.958585] env[62368]: DEBUG oslo.service.loopingcall [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.958779] env[62368]: DEBUG nova.compute.manager [-] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.958874] env[62368]: DEBUG nova.network.neutron [-] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 755.008991] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198300, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.072853] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.112900] env[62368]: DEBUG nova.network.neutron [-] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.130970] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198301, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063203} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.131930] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.132720] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dca5b6-57db-4723-af00-55d6257f0f04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.140377] env[62368]: DEBUG nova.compute.utils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.162882] env[62368]: DEBUG nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.162882] env[62368]: DEBUG nova.network.neutron [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.175560] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d/4fe7b1a2-f894-4131-9456-fb4df8a4532d.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.180148] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f464f69e-e7ea-47a0-b536-a8dfad3a0511 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.206382] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 755.206382] env[62368]: value = "task-1198305" [ 755.206382] env[62368]: _type = "Task" [ 755.206382] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.215548] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198305, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.220863] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.221046] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 3f36000e-b93d-4dda-ac39-b8459203c227 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 755.221186] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 755.221314] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 4dc7b6b4-1bf5-4195-bb93-14756f8f9986 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 755.221432] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.221548] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b393f654-0611-493e-a89f-445dfd3f8c13 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.221663] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 085ebe93-aa24-4626-94fe-241c4297e4db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.221833] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 13765305-2e55-4ee8-9a6f-4ae5ee724367 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.221887] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 4fe7b1a2-f894-4131-9456-fb4df8a4532d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.221983] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 417f000f-cf23-404d-877c-45990d1a7c77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.222102] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.277504] env[62368]: DEBUG nova.policy [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7dceb8c8f8b4c4dbdde4d097009d429', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b70fcc6664f47d7b55447210851c4cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.403034] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198304, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.514669] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198300, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.559011] env[62368]: DEBUG nova.compute.manager [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Received event network-changed-6e4af274-eb1c-4a81-b1f6-dc415b7be3d5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.559323] env[62368]: DEBUG nova.compute.manager [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Refreshing instance network info cache due to event network-changed-6e4af274-eb1c-4a81-b1f6-dc415b7be3d5. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 755.559597] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] Acquiring lock "refresh_cache-417f000f-cf23-404d-877c-45990d1a7c77" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.559791] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] Acquired lock "refresh_cache-417f000f-cf23-404d-877c-45990d1a7c77" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.560015] env[62368]: DEBUG nova.network.neutron [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Refreshing network info cache for port 6e4af274-eb1c-4a81-b1f6-dc415b7be3d5 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 755.618208] env[62368]: INFO nova.compute.manager [-] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Took 1.32 seconds to deallocate network for instance. [ 755.626700] env[62368]: DEBUG nova.network.neutron [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Successfully created port: b829cb49-cde3-4f16-8e08-6f56dff38e09 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.650178] env[62368]: DEBUG nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.720515] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198305, 'name': ReconfigVM_Task, 'duration_secs': 0.368698} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.721067] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d/4fe7b1a2-f894-4131-9456-fb4df8a4532d.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 755.721781] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16ed4f7d-ebb6-4acd-b120-b3650f2f25d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.724411] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.730155] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 755.730155] env[62368]: value = "task-1198306" [ 755.730155] env[62368]: _type = "Task" [ 755.730155] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.741512] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198306, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.901572] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198304, 'name': CreateVM_Task, 'duration_secs': 0.65489} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.901743] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 755.902460] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.902625] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.902947] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 755.903212] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57a6731a-6c76-44f7-9018-c9da53374a59 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.909277] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 755.909277] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52eda148-0535-8212-fd96-8f636a097642" [ 755.909277] env[62368]: _type = "Task" [ 755.909277] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.917926] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52eda148-0535-8212-fd96-8f636a097642, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.010222] env[62368]: DEBUG oslo_vmware.api [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198300, 'name': PowerOnVM_Task, 'duration_secs': 1.068591} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.010670] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 756.010715] env[62368]: INFO nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Took 8.94 seconds to spawn the instance on the hypervisor. [ 756.010893] env[62368]: DEBUG nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 756.011683] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc734de-1b85-42d9-9534-b2a1e4023f3f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.063803] env[62368]: DEBUG nova.network.neutron [-] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.126278] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.226916] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 300c6c32-b365-40c2-bc7d-6ce028f10476 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.241401] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198306, 'name': Rename_Task, 'duration_secs': 0.147727} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.242296] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 756.242705] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b98b1ba6-1178-4f40-b4e7-72e51d3a6393 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.254856] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 756.254856] env[62368]: value = "task-1198307" [ 756.254856] env[62368]: _type = "Task" [ 756.254856] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.263603] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198307, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.366281] env[62368]: DEBUG nova.network.neutron [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Updated VIF entry in instance network info cache for port 6e4af274-eb1c-4a81-b1f6-dc415b7be3d5. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 756.366741] env[62368]: DEBUG nova.network.neutron [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Updating instance_info_cache with network_info: [{"id": "6e4af274-eb1c-4a81-b1f6-dc415b7be3d5", "address": "fa:16:3e:2d:d5:37", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e4af274-eb", "ovs_interfaceid": "6e4af274-eb1c-4a81-b1f6-dc415b7be3d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.422680] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52eda148-0535-8212-fd96-8f636a097642, 'name': SearchDatastore_Task, 'duration_secs': 0.010327} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.423025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.423279] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 756.423581] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.423728] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.423909] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 756.424181] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1354017-dca1-4b50-ba2b-a93511caa677 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.433798] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 756.433994] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 756.435225] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38006f93-1bb0-4ba6-adaa-39c34693afa0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.440835] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 756.440835] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]523fdbd8-cfba-7e02-d9fb-606cfd00ab6e" [ 756.440835] env[62368]: _type = "Task" [ 756.440835] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.243271] env[62368]: INFO nova.compute.manager [-] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Took 2.28 seconds to deallocate network for instance. [ 757.243271] env[62368]: DEBUG nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.245448] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 3f1af54b-392f-432a-9ffa-a133da428f94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.248397] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] Releasing lock "refresh_cache-417f000f-cf23-404d-877c-45990d1a7c77" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.248397] env[62368]: DEBUG nova.compute.manager [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Received event network-changed-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.248397] env[62368]: DEBUG nova.compute.manager [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Refreshing instance network info cache due to event network-changed-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.248397] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] Acquiring lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.248397] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] Acquired lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.248679] env[62368]: DEBUG nova.network.neutron [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Refreshing network info cache for port 3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.259032] env[62368]: INFO nova.compute.manager [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Took 34.69 seconds to build instance. [ 757.265665] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523fdbd8-cfba-7e02-d9fb-606cfd00ab6e, 'name': SearchDatastore_Task, 'duration_secs': 0.015363} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.269670] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14315728-ba95-4d81-99f7-19284f2e060d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.273036] env[62368]: DEBUG oslo_vmware.api [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198307, 'name': PowerOnVM_Task, 'duration_secs': 0.561837} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.275654] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 757.275654] env[62368]: INFO nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Took 7.69 seconds to spawn the instance on the hypervisor. [ 757.275654] env[62368]: DEBUG nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.275654] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4fed9c-6d72-41ae-bd8e-405486cc7568 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.279446] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 757.279446] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5226d355-1b5f-23b1-7196-5bf9ccecf556" [ 757.279446] env[62368]: _type = "Task" [ 757.279446] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.293434] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5226d355-1b5f-23b1-7196-5bf9ccecf556, 'name': SearchDatastore_Task, 'duration_secs': 0.011579} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.293434] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.293682] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 417f000f-cf23-404d-877c-45990d1a7c77/417f000f-cf23-404d-877c-45990d1a7c77.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 757.293974] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4eff641-73c3-4a5f-b481-70206ec96675 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.305131] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 757.305131] env[62368]: value = "task-1198308" [ 757.305131] env[62368]: _type = "Task" [ 757.305131] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.316845] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.322529] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.322811] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.323030] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.323318] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.323532] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.323723] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.323967] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.324794] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.324794] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.324794] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.324794] env[62368]: DEBUG nova.virt.hardware [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.325857] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad24758f-20ec-4463-976c-a4058e23f3dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.335629] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd83d726-cefb-4946-929e-c29668def2c3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.595600] env[62368]: DEBUG nova.compute.manager [req-ae79b9d0-5aa6-45ea-8202-fe8e6279b261 req-c6b38a53-f3a5-42eb-927e-67655989f62d service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Received event network-changed-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.595893] env[62368]: DEBUG nova.compute.manager [req-ae79b9d0-5aa6-45ea-8202-fe8e6279b261 req-c6b38a53-f3a5-42eb-927e-67655989f62d service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Refreshing instance network info cache due to event network-changed-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.596129] env[62368]: DEBUG oslo_concurrency.lockutils [req-ae79b9d0-5aa6-45ea-8202-fe8e6279b261 req-c6b38a53-f3a5-42eb-927e-67655989f62d service nova] Acquiring lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.751568] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 67d8c43a-d3af-407b-847a-1c2c8c53dbfb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.765521] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.767310] env[62368]: DEBUG oslo_concurrency.lockutils [None req-47dc897d-53b8-4513-979c-a3eb12c15d71 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.251s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.805159] env[62368]: INFO nova.compute.manager [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Took 30.79 seconds to build instance. [ 757.820225] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198308, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.971224] env[62368]: DEBUG nova.compute.manager [req-362939f2-7d62-4234-a798-a387e0e2f42b req-e0e698a4-d6ff-427d-a7ac-0bd0e23b89f0 service nova] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Received event network-vif-deleted-4b2bdc5d-6176-4415-a60e-9519b40bd055 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.003300] env[62368]: DEBUG nova.network.neutron [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Successfully updated port: b829cb49-cde3-4f16-8e08-6f56dff38e09 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 758.200992] env[62368]: DEBUG nova.network.neutron [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Updated VIF entry in instance network info cache for port 3c5d30b5-42b0-4dc4-81a6-b9733f7e0797. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 758.201485] env[62368]: DEBUG nova.network.neutron [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Updating instance_info_cache with network_info: [{"id": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "address": "fa:16:3e:b9:01:76", "network": {"id": "8c5de397-da1e-4836-a5f8-cd85f07a1239", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-831639205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b12e7fa4eec4f3b975651f2fc46c8fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c5d30b5-42", "ovs_interfaceid": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.255321] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 0f1bff84-1fb9-471a-b685-7d527b376b0a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.270313] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 758.311859] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a24fd7c-9604-42f7-b541-f2821b62aa7f tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.607s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.318467] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198308, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523874} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.318940] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 417f000f-cf23-404d-877c-45990d1a7c77/417f000f-cf23-404d-877c-45990d1a7c77.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 758.318940] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 758.319207] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-618268f1-c42d-4eb4-b9c1-99d2faa3a02e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.327219] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 758.327219] env[62368]: value = "task-1198309" [ 758.327219] env[62368]: _type = "Task" [ 758.327219] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.340582] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198309, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.512621] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.512621] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.512621] env[62368]: DEBUG nova.network.neutron [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.704821] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] Releasing lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.705106] env[62368]: DEBUG nova.compute.manager [req-1ce2b962-1d38-48c7-a708-4340868261f2 req-ec8cae43-9737-4d91-9311-5806be25c1ff service nova] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Received event network-vif-deleted-6db52e60-d5f1-4028-b5ff-ffd3bc376cfb {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.705559] env[62368]: DEBUG oslo_concurrency.lockutils [req-ae79b9d0-5aa6-45ea-8202-fe8e6279b261 req-c6b38a53-f3a5-42eb-927e-67655989f62d service nova] Acquired lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.705759] env[62368]: DEBUG nova.network.neutron [req-ae79b9d0-5aa6-45ea-8202-fe8e6279b261 req-c6b38a53-f3a5-42eb-927e-67655989f62d service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Refreshing network info cache for port 3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 758.758441] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.797381] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.815029] env[62368]: DEBUG nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 758.839051] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198309, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072839} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.839324] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.840163] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b28e1a06-18d9-4914-abd1-1cf256559513 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.864214] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 417f000f-cf23-404d-877c-45990d1a7c77/417f000f-cf23-404d-877c-45990d1a7c77.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.865315] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcd40f5b-35cc-45b9-9cdb-f1c3949590b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.892997] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 758.892997] env[62368]: value = "task-1198310" [ 758.892997] env[62368]: _type = "Task" [ 758.892997] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.899189] env[62368]: INFO nova.compute.manager [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Rescuing [ 758.899447] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.899597] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquired lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.899757] env[62368]: DEBUG nova.network.neutron [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.905129] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198310, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.073924] env[62368]: DEBUG nova.network.neutron [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.261693] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 87dc392f-7bba-44bb-ba38-2fd4e1c6129e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.332699] env[62368]: DEBUG nova.network.neutron [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Updating instance_info_cache with network_info: [{"id": "b829cb49-cde3-4f16-8e08-6f56dff38e09", "address": "fa:16:3e:68:ed:ea", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb829cb49-cd", "ovs_interfaceid": "b829cb49-cde3-4f16-8e08-6f56dff38e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.346268] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.387645] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquiring lock "085ebe93-aa24-4626-94fe-241c4297e4db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.387645] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "085ebe93-aa24-4626-94fe-241c4297e4db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.387863] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquiring lock "085ebe93-aa24-4626-94fe-241c4297e4db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.388019] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "085ebe93-aa24-4626-94fe-241c4297e4db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.388202] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "085ebe93-aa24-4626-94fe-241c4297e4db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.395037] env[62368]: INFO nova.compute.manager [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Terminating instance [ 759.398690] env[62368]: DEBUG nova.compute.manager [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.398955] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 759.399843] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37553ec-6660-4d62-9e31-eb39f0c83316 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.415176] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198310, 'name': ReconfigVM_Task, 'duration_secs': 0.511657} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.417342] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 417f000f-cf23-404d-877c-45990d1a7c77/417f000f-cf23-404d-877c-45990d1a7c77.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 759.420069] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 759.420566] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45db8a29-76e4-4508-8038-bfed6d28faf7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.423411] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07733012-7c6a-4f37-804e-e28319718c58 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.433571] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 759.433571] env[62368]: value = "task-1198311" [ 759.433571] env[62368]: _type = "Task" [ 759.433571] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.435429] env[62368]: DEBUG oslo_vmware.api [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 759.435429] env[62368]: value = "task-1198312" [ 759.435429] env[62368]: _type = "Task" [ 759.435429] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.451101] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198311, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.457170] env[62368]: DEBUG oslo_vmware.api [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198312, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.487847] env[62368]: DEBUG nova.network.neutron [req-ae79b9d0-5aa6-45ea-8202-fe8e6279b261 req-c6b38a53-f3a5-42eb-927e-67655989f62d service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Updated VIF entry in instance network info cache for port 3c5d30b5-42b0-4dc4-81a6-b9733f7e0797. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 759.488278] env[62368]: DEBUG nova.network.neutron [req-ae79b9d0-5aa6-45ea-8202-fe8e6279b261 req-c6b38a53-f3a5-42eb-927e-67655989f62d service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Updating instance_info_cache with network_info: [{"id": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "address": "fa:16:3e:b9:01:76", "network": {"id": "8c5de397-da1e-4836-a5f8-cd85f07a1239", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-831639205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b12e7fa4eec4f3b975651f2fc46c8fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c5d30b5-42", "ovs_interfaceid": "3c5d30b5-42b0-4dc4-81a6-b9733f7e0797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.661915] env[62368]: DEBUG nova.compute.manager [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Received event network-vif-plugged-b829cb49-cde3-4f16-8e08-6f56dff38e09 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.663099] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Acquiring lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.663099] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.663099] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.663099] env[62368]: DEBUG nova.compute.manager [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] No waiting events found dispatching network-vif-plugged-b829cb49-cde3-4f16-8e08-6f56dff38e09 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 759.663099] env[62368]: WARNING nova.compute.manager [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Received unexpected event network-vif-plugged-b829cb49-cde3-4f16-8e08-6f56dff38e09 for instance with vm_state building and task_state spawning. [ 759.663234] env[62368]: DEBUG nova.compute.manager [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Received event network-changed-11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.663234] env[62368]: DEBUG nova.compute.manager [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Refreshing instance network info cache due to event network-changed-11ce4c60-0df0-4c43-bbe4-6cc53717309b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.663423] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Acquiring lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.663566] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Acquired lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.663800] env[62368]: DEBUG nova.network.neutron [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Refreshing network info cache for port 11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 759.764418] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a0cea538-b162-4504-ac34-803a2d5a8071 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.775816] env[62368]: DEBUG nova.network.neutron [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updating instance_info_cache with network_info: [{"id": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "address": "fa:16:3e:fc:46:9c", "network": {"id": "d71a42e9-1e1c-4abd-b585-2362ba1b8cfb", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1629701000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3b87e3b7d1d947b79f4834e4e0fe0395", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280f8ef3-0f", "ovs_interfaceid": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.841013] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.841013] env[62368]: DEBUG nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Instance network_info: |[{"id": "b829cb49-cde3-4f16-8e08-6f56dff38e09", "address": "fa:16:3e:68:ed:ea", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb829cb49-cd", "ovs_interfaceid": "b829cb49-cde3-4f16-8e08-6f56dff38e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 759.841242] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:ed:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b829cb49-cde3-4f16-8e08-6f56dff38e09', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.850531] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating folder: Project (9b70fcc6664f47d7b55447210851c4cc). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 759.851487] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-546c4644-35aa-4544-a7d0-64be5f3d0a3a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.864230] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Created folder: Project (9b70fcc6664f47d7b55447210851c4cc) in parent group-v259706. [ 759.864230] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating folder: Instances. Parent ref: group-v259761. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 759.864230] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5904235-d68c-4fef-9a29-33f0f3e7e2b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.874972] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Created folder: Instances in parent group-v259761. [ 759.875262] env[62368]: DEBUG oslo.service.loopingcall [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.875465] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 759.875678] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a403baf-3227-46db-ad05-ebe76477d4c5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.898140] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 759.898140] env[62368]: value = "task-1198315" [ 759.898140] env[62368]: _type = "Task" [ 759.898140] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.907827] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198315, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.950426] env[62368]: DEBUG oslo_vmware.api [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198312, 'name': PowerOffVM_Task, 'duration_secs': 0.21363} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.955061] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 759.955172] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 759.956735] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198311, 'name': Rename_Task, 'duration_secs': 0.225492} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.956735] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-41097766-6e98-4f48-9d34-0050ff5a777b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.957538] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 759.957538] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12dc4101-59ec-4711-a5a2-519d2bebae9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.965355] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 759.965355] env[62368]: value = "task-1198317" [ 759.965355] env[62368]: _type = "Task" [ 759.965355] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.974788] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.996400] env[62368]: DEBUG oslo_concurrency.lockutils [req-ae79b9d0-5aa6-45ea-8202-fe8e6279b261 req-c6b38a53-f3a5-42eb-927e-67655989f62d service nova] Releasing lock "refresh_cache-085ebe93-aa24-4626-94fe-241c4297e4db" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.040191] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 760.040887] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 760.040887] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Deleting the datastore file [datastore1] 085ebe93-aa24-4626-94fe-241c4297e4db {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 760.041173] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58ec8b4d-0116-40ad-8303-889fbc3586ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.050992] env[62368]: DEBUG oslo_vmware.api [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for the task: (returnval){ [ 760.050992] env[62368]: value = "task-1198318" [ 760.050992] env[62368]: _type = "Task" [ 760.050992] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.061281] env[62368]: DEBUG oslo_vmware.api [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198318, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.269763] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b79f0e79-9e3a-47c7-9949-8743601ec6c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.284415] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Releasing lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.409785] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198315, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.475578] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198317, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.561571] env[62368]: DEBUG oslo_vmware.api [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Task: {'id': task-1198318, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188661} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.561891] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 760.562119] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 760.562337] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 760.562542] env[62368]: INFO nova.compute.manager [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Took 1.16 seconds to destroy the instance on the hypervisor. [ 760.562833] env[62368]: DEBUG oslo.service.loopingcall [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.567017] env[62368]: DEBUG nova.compute.manager [-] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.567017] env[62368]: DEBUG nova.network.neutron [-] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 760.601689] env[62368]: DEBUG nova.network.neutron [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updated VIF entry in instance network info cache for port 11ce4c60-0df0-4c43-bbe4-6cc53717309b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 760.602042] env[62368]: DEBUG nova.network.neutron [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.774341] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 54083854-e314-44a7-b4b0-fbef5fa6b1ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.827472] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 760.828084] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9368e5d-b491-45d6-8091-f767cd863daa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.840216] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 760.840216] env[62368]: value = "task-1198319" [ 760.840216] env[62368]: _type = "Task" [ 760.840216] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.851025] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.909558] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198315, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.978556] env[62368]: DEBUG oslo_vmware.api [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198317, 'name': PowerOnVM_Task, 'duration_secs': 0.991136} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.978878] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 760.979103] env[62368]: INFO nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Took 8.89 seconds to spawn the instance on the hypervisor. [ 760.979288] env[62368]: DEBUG nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.980545] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12025c6e-f81d-4a2f-b2d2-4b3a41d20ffa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.105521] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Releasing lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.105924] env[62368]: DEBUG nova.compute.manager [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Received event network-changed-b829cb49-cde3-4f16-8e08-6f56dff38e09 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.106126] env[62368]: DEBUG nova.compute.manager [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Refreshing instance network info cache due to event network-changed-b829cb49-cde3-4f16-8e08-6f56dff38e09. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.106455] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Acquiring lock "refresh_cache-af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.106605] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Acquired lock "refresh_cache-af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.106769] env[62368]: DEBUG nova.network.neutron [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Refreshing network info cache for port b829cb49-cde3-4f16-8e08-6f56dff38e09 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 761.282465] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance eea21546-fbbf-4440-829c-8583c4ccabb6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.282465] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 761.282465] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 761.348706] env[62368]: DEBUG nova.network.neutron [-] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.352735] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.412560] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198315, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.505196] env[62368]: INFO nova.compute.manager [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Took 32.78 seconds to build instance. [ 761.664588] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183ed32f-5220-4954-bdb5-be8fccf6dc73 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.672468] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468a7136-718c-4ae6-b8e6-b756ddd94be8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.721224] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04deb4e-43b2-4c7e-8a82-7790d39e568a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.732524] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1c051c-edb4-4150-874f-245e50ee66d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.746843] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.842922] env[62368]: DEBUG nova.compute.manager [req-3f252940-5ad4-4c07-abb2-d65f84bee981 req-e95b5897-e8e9-4db5-9972-417163704c81 service nova] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Received event network-vif-deleted-3c5d30b5-42b0-4dc4-81a6-b9733f7e0797 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.852605] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198319, 'name': PowerOffVM_Task, 'duration_secs': 0.726473} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.853900] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 761.853900] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d61ebb7-20b1-4630-8eb9-41103c6e273a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.856784] env[62368]: INFO nova.compute.manager [-] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Took 1.29 seconds to deallocate network for instance. [ 761.882040] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90a07a2-c09c-44ba-950b-37585780aea3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.915563] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198315, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.918945] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 761.919558] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4978cdfa-a1e2-481d-bdae-ca0d10de6add {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.928513] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 761.928513] env[62368]: value = "task-1198320" [ 761.928513] env[62368]: _type = "Task" [ 761.928513] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.932417] env[62368]: DEBUG nova.network.neutron [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Updated VIF entry in instance network info cache for port b829cb49-cde3-4f16-8e08-6f56dff38e09. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 761.932842] env[62368]: DEBUG nova.network.neutron [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Updating instance_info_cache with network_info: [{"id": "b829cb49-cde3-4f16-8e08-6f56dff38e09", "address": "fa:16:3e:68:ed:ea", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb829cb49-cd", "ovs_interfaceid": "b829cb49-cde3-4f16-8e08-6f56dff38e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.941032] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 761.941136] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 761.941331] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.941489] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.941662] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 761.942209] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49e075b0-460c-400b-9925-63d66d6e67c3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.952527] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 761.952714] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 761.954034] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe4b6458-353a-4199-8ca9-872d705ed077 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.959465] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 761.959465] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5219b0cf-aab9-03fa-dbb3-a16f113a44da" [ 761.959465] env[62368]: _type = "Task" [ 761.959465] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.968018] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5219b0cf-aab9-03fa-dbb3-a16f113a44da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.007164] env[62368]: DEBUG oslo_concurrency.lockutils [None req-533bf0a9-fa77-4487-90ae-a796f145bd35 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "417f000f-cf23-404d-877c-45990d1a7c77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.844s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.222127] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.222409] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.252032] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.378326] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.414274] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198315, 'name': CreateVM_Task, 'duration_secs': 2.271832} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.414426] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 762.415147] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.415329] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.415643] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.415902] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ba6b021-5840-49b2-977f-1096d17e1229 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.420758] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 762.420758] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52cf717c-838a-e8ef-f323-65102f3cf59c" [ 762.420758] env[62368]: _type = "Task" [ 762.420758] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.428646] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cf717c-838a-e8ef-f323-65102f3cf59c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.435318] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adcf589-5f9c-46a9-bcba-89ad6053ec83 req-bb2016d2-573d-4c4b-8c0d-c5c870862b10 service nova] Releasing lock "refresh_cache-af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.468701] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5219b0cf-aab9-03fa-dbb3-a16f113a44da, 'name': SearchDatastore_Task, 'duration_secs': 0.009649} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.469473] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3fc2850-74b0-4ef4-8a75-b305a420d4f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.474512] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 762.474512] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52633bb1-22b5-a3a2-74d2-a98d9c77f249" [ 762.474512] env[62368]: _type = "Task" [ 762.474512] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.484570] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52633bb1-22b5-a3a2-74d2-a98d9c77f249, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.511137] env[62368]: DEBUG nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.756895] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 762.757226] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.120s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.757530] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.974s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.759152] env[62368]: INFO nova.compute.claims [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.762254] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.762434] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Cleaning up deleted instances {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 762.931662] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cf717c-838a-e8ef-f323-65102f3cf59c, 'name': SearchDatastore_Task, 'duration_secs': 0.009413} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.931976] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.932230] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.932462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.932607] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.932782] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.933048] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44735041-8c21-4fe0-9afb-feac2916e491 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.942546] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.942727] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 762.943483] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ea60cc4-1167-436d-b1c4-ec4f954487cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.948769] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 762.948769] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f6e649-1e96-0069-c37b-9c4fdbdd70d0" [ 762.948769] env[62368]: _type = "Task" [ 762.948769] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.956380] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f6e649-1e96-0069-c37b-9c4fdbdd70d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.984041] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52633bb1-22b5-a3a2-74d2-a98d9c77f249, 'name': SearchDatastore_Task, 'duration_secs': 0.009154} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.984298] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.984605] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. {{(pid=62368) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 762.984873] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fcaee372-93a9-47f3-a849-40555f68152e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.988874] env[62368]: DEBUG nova.compute.manager [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 762.989607] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4d3c84-b2a9-4282-a502-6f2b6d4f9f05 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.999172] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 762.999172] env[62368]: value = "task-1198321" [ 762.999172] env[62368]: _type = "Task" [ 762.999172] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.007383] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.036678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.272862] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] There are 5 instances to clean {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 763.272862] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a5c638d8-f761-497d-bde2-39ea90d8d151] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 763.462017] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f6e649-1e96-0069-c37b-9c4fdbdd70d0, 'name': SearchDatastore_Task, 'duration_secs': 0.009522} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.466019] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b99cc8b-c570-4fe5-8dc6-e4447009da3d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.469049] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 763.469049] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52845ecd-a199-4681-01a1-15e544045f07" [ 763.469049] env[62368]: _type = "Task" [ 763.469049] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.478400] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52845ecd-a199-4681-01a1-15e544045f07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.501331] env[62368]: INFO nova.compute.manager [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] instance snapshotting [ 763.507649] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0729abe5-7642-493a-b7d9-72a30844525b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.516976] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453081} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.531443] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. [ 763.533433] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ac3ed8-6f9b-4a34-9c0b-faa3d103ec82 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.536841] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7221df6-4b8a-451d-9392-5d81912ffced {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.570974] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 763.571690] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2248f5e1-80aa-4081-b173-b80eaded4956 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.595317] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 763.595317] env[62368]: value = "task-1198322" [ 763.595317] env[62368]: _type = "Task" [ 763.595317] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.606785] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198322, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.778017] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 38291806-d0ae-47a5-bc6e-3d0320b2bd50] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 763.980931] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52845ecd-a199-4681-01a1-15e544045f07, 'name': SearchDatastore_Task, 'duration_secs': 0.014203} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.981052] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.981333] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03/af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 763.984047] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef47e6a0-b869-4988-bee8-afbc8a1e2db6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.992313] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 763.992313] env[62368]: value = "task-1198323" [ 763.992313] env[62368]: _type = "Task" [ 763.992313] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.001876] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.073172] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 764.073596] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-eefef180-03d5-47cc-a9b7-73467e2e8070 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.083740] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 764.083740] env[62368]: value = "task-1198324" [ 764.083740] env[62368]: _type = "Task" [ 764.083740] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.093752] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198324, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.095676] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff06424-fddb-4543-b13a-949f87f693a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.109910] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7deb5cf4-ce3d-4db1-a6be-865624287203 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.113051] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198322, 'name': ReconfigVM_Task, 'duration_secs': 0.297364} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.113319] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 764.114510] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239829c5-0177-4ee3-92db-edeb76251208 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.142899] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf57c32-0fe1-4c51-87a6-eae9402ccfa2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.168587] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8111ef8d-e8dc-4085-acae-469bf7e3032a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.184744] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44884946-0eee-4419-8014-0ac8a9ff1a64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.192631] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 764.192631] env[62368]: value = "task-1198325" [ 764.192631] env[62368]: _type = "Task" [ 764.192631] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.206881] env[62368]: DEBUG nova.compute.provider_tree [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.218163] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.281473] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 4fde6065-41a7-4a95-96df-e442762dad24] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 764.502656] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497327} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.502960] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03/af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 764.503163] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 764.503423] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af86b2b0-f8c5-44ab-b897-151f1315ef59 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.510097] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 764.510097] env[62368]: value = "task-1198326" [ 764.510097] env[62368]: _type = "Task" [ 764.510097] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.519076] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198326, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.593271] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198324, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.704087] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198325, 'name': ReconfigVM_Task, 'duration_secs': 0.200046} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.704408] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 764.704736] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ced9de59-1ea2-4162-b193-8ad9f34590eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.709634] env[62368]: DEBUG nova.scheduler.client.report [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.714227] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 764.714227] env[62368]: value = "task-1198327" [ 764.714227] env[62368]: _type = "Task" [ 764.714227] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.723315] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198327, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.784624] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: d22aff27-739f-4b0d-a0e4-b6316c252c3a] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 765.024492] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198326, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059735} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.024798] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.026057] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6935992f-74d6-4deb-b5a9-a057682ca18b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.063948] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03/af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.064441] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81c2a4e0-9fa6-4efe-b45d-2e2d24a1d55b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.105698] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198324, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.107738] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 765.107738] env[62368]: value = "task-1198328" [ 765.107738] env[62368]: _type = "Task" [ 765.107738] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.118937] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198328, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.217299] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.218102] env[62368]: DEBUG nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 765.221928] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.740s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.224044] env[62368]: INFO nova.compute.claims [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.240483] env[62368]: DEBUG oslo_vmware.api [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198327, 'name': PowerOnVM_Task, 'duration_secs': 0.408431} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.240763] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 765.243696] env[62368]: DEBUG nova.compute.manager [None req-9a858e87-5789-4d2e-b660-54fff9ac3a02 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 765.244565] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30450819-31d5-4bfd-bfce-594b4fc41681 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.287638] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 0eb2a772-7635-481a-ae2c-9f23669dd175] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 765.603906] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198324, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.617560] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.736371] env[62368]: DEBUG nova.compute.utils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.737778] env[62368]: DEBUG nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.737954] env[62368]: DEBUG nova.network.neutron [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 765.792990] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.793463] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Cleaning up deleted instances with incomplete migration {{(pid=62368) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 765.801177] env[62368]: DEBUG nova.policy [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ec6fb90433943ca9711d4a7a5e40bef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd4cd004b0d54bb49a88722fd1312249', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 766.108781] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198324, 'name': CreateSnapshot_Task, 'duration_secs': 1.610015} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.114099] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 766.114709] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e3b7fd-238f-4303-a3e5-360e50199af5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.123199] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198328, 'name': ReconfigVM_Task, 'duration_secs': 0.850744} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.124022] env[62368]: DEBUG nova.network.neutron [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Successfully created port: 4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.129843] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Reconfigured VM instance instance-00000035 to attach disk [datastore1] af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03/af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.132212] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f5974da-8762-4b6e-9626-923680230762 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.144903] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 766.144903] env[62368]: value = "task-1198329" [ 766.144903] env[62368]: _type = "Task" [ 766.144903] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.168555] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198329, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.241747] env[62368]: DEBUG nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.296674] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.568127] env[62368]: DEBUG nova.compute.manager [req-66da4977-fcfc-445f-8d01-9a683366aa0a req-ca22c647-43da-4332-bf25-de62333298bd service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Received event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.568127] env[62368]: DEBUG nova.compute.manager [req-66da4977-fcfc-445f-8d01-9a683366aa0a req-ca22c647-43da-4332-bf25-de62333298bd service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing instance network info cache due to event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 766.568127] env[62368]: DEBUG oslo_concurrency.lockutils [req-66da4977-fcfc-445f-8d01-9a683366aa0a req-ca22c647-43da-4332-bf25-de62333298bd service nova] Acquiring lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.568127] env[62368]: DEBUG oslo_concurrency.lockutils [req-66da4977-fcfc-445f-8d01-9a683366aa0a req-ca22c647-43da-4332-bf25-de62333298bd service nova] Acquired lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.568416] env[62368]: DEBUG nova.network.neutron [req-66da4977-fcfc-445f-8d01-9a683366aa0a req-ca22c647-43da-4332-bf25-de62333298bd service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 766.591039] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2371a2bd-12bd-4b5b-930a-cec8e1cb5e95 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.600682] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4c6346-7799-4d45-997e-8ac28392dc01 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.633677] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41377232-6be4-4129-b307-377144689eab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.642864] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 766.643213] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7c844072-df14-4199-8454-6765932ed744 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.655777] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bc7160-d8e6-4f85-967b-88b32b7a375c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.663588] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 766.663588] env[62368]: value = "task-1198330" [ 766.663588] env[62368]: _type = "Task" [ 766.663588] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.663834] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198329, 'name': Rename_Task, 'duration_secs': 0.164219} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.664540] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 766.667932] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa4df219-70b7-4445-9012-b0c9dd36799b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.677540] env[62368]: DEBUG nova.compute.provider_tree [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.685123] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198330, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.686586] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 766.686586] env[62368]: value = "task-1198331" [ 766.686586] env[62368]: _type = "Task" [ 766.686586] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.700550] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198331, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.956457] env[62368]: DEBUG nova.compute.manager [req-30ff161d-8775-4548-b5f3-b76473fed80c req-31cab586-b23c-4e9c-a649-5f2b398b9b87 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Received event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.956693] env[62368]: DEBUG nova.compute.manager [req-30ff161d-8775-4548-b5f3-b76473fed80c req-31cab586-b23c-4e9c-a649-5f2b398b9b87 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing instance network info cache due to event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 766.956923] env[62368]: DEBUG oslo_concurrency.lockutils [req-30ff161d-8775-4548-b5f3-b76473fed80c req-31cab586-b23c-4e9c-a649-5f2b398b9b87 service nova] Acquiring lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.175424] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198330, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.182580] env[62368]: DEBUG nova.scheduler.client.report [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.200373] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198331, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.259980] env[62368]: DEBUG nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.294269] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.294570] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.294729] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.294913] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.295072] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.295223] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.295532] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.295749] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.295919] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.296101] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.296277] env[62368]: DEBUG nova.virt.hardware [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.297146] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0188dcf-d121-436e-bd53-5ba961ff772a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.305879] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fc9a84-d5bd-4d65-87e4-6c41c7f0c54d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.433028] env[62368]: DEBUG nova.network.neutron [req-66da4977-fcfc-445f-8d01-9a683366aa0a req-ca22c647-43da-4332-bf25-de62333298bd service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updated VIF entry in instance network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 767.433028] env[62368]: DEBUG nova.network.neutron [req-66da4977-fcfc-445f-8d01-9a683366aa0a req-ca22c647-43da-4332-bf25-de62333298bd service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updating instance_info_cache with network_info: [{"id": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "address": "fa:16:3e:fc:46:9c", "network": {"id": "d71a42e9-1e1c-4abd-b585-2362ba1b8cfb", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1629701000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3b87e3b7d1d947b79f4834e4e0fe0395", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280f8ef3-0f", "ovs_interfaceid": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.676709] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198330, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.687734] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.688460] env[62368]: DEBUG nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 767.691121] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.833s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.691324] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.693409] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.621s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.695788] env[62368]: INFO nova.compute.claims [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.709050] env[62368]: DEBUG oslo_vmware.api [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198331, 'name': PowerOnVM_Task, 'duration_secs': 0.555489} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.709318] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 767.709618] env[62368]: INFO nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Took 10.47 seconds to spawn the instance on the hypervisor. [ 767.709817] env[62368]: DEBUG nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.710909] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce49c4e-7a56-4481-be80-1b682cbf1873 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.746646] env[62368]: INFO nova.scheduler.client.report [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Deleted allocations for instance 3f36000e-b93d-4dda-ac39-b8459203c227 [ 767.933079] env[62368]: DEBUG oslo_concurrency.lockutils [req-66da4977-fcfc-445f-8d01-9a683366aa0a req-ca22c647-43da-4332-bf25-de62333298bd service nova] Releasing lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.934513] env[62368]: DEBUG oslo_concurrency.lockutils [req-30ff161d-8775-4548-b5f3-b76473fed80c req-31cab586-b23c-4e9c-a649-5f2b398b9b87 service nova] Acquired lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.934870] env[62368]: DEBUG nova.network.neutron [req-30ff161d-8775-4548-b5f3-b76473fed80c req-31cab586-b23c-4e9c-a649-5f2b398b9b87 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 767.998016] env[62368]: DEBUG nova.network.neutron [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Successfully updated port: 4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 768.177633] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198330, 'name': CloneVM_Task} progress is 95%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.203055] env[62368]: DEBUG nova.compute.utils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.209620] env[62368]: DEBUG nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.209776] env[62368]: DEBUG nova.network.neutron [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 768.233340] env[62368]: INFO nova.compute.manager [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Took 31.74 seconds to build instance. [ 768.257070] env[62368]: DEBUG oslo_concurrency.lockutils [None req-539123f0-9140-49de-a2ef-aca377acf04c tempest-ServerTagsTestJSON-1285862186 tempest-ServerTagsTestJSON-1285862186-project-member] Lock "3f36000e-b93d-4dda-ac39-b8459203c227" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.335s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.302767] env[62368]: DEBUG nova.policy [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72595ab2bbf94ec5ac63a5b5351ae798', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ff43f079c154fc6982d0cd583c6f789', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 768.501788] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.501868] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquired lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.502036] env[62368]: DEBUG nova.network.neutron [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 768.676821] env[62368]: DEBUG nova.compute.manager [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Received event network-vif-plugged-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.677127] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] Acquiring lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.677303] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.677504] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.677697] env[62368]: DEBUG nova.compute.manager [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] No waiting events found dispatching network-vif-plugged-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 768.677864] env[62368]: WARNING nova.compute.manager [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Received unexpected event network-vif-plugged-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c for instance with vm_state building and task_state spawning. [ 768.678028] env[62368]: DEBUG nova.compute.manager [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Received event network-changed-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.678244] env[62368]: DEBUG nova.compute.manager [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Refreshing instance network info cache due to event network-changed-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.678339] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] Acquiring lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.682373] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198330, 'name': CloneVM_Task, 'duration_secs': 1.69} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.682633] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Created linked-clone VM from snapshot [ 768.683402] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125f3e88-ec14-4f24-a2f1-1c4e600c970e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.692475] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Uploading image 6281ea63-c477-4e65-8fac-7b43765b0423 {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 768.707776] env[62368]: DEBUG nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 768.724190] env[62368]: DEBUG nova.network.neutron [req-30ff161d-8775-4548-b5f3-b76473fed80c req-31cab586-b23c-4e9c-a649-5f2b398b9b87 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updated VIF entry in instance network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 768.728113] env[62368]: DEBUG nova.network.neutron [req-30ff161d-8775-4548-b5f3-b76473fed80c req-31cab586-b23c-4e9c-a649-5f2b398b9b87 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updating instance_info_cache with network_info: [{"id": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "address": "fa:16:3e:fc:46:9c", "network": {"id": "d71a42e9-1e1c-4abd-b585-2362ba1b8cfb", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1629701000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3b87e3b7d1d947b79f4834e4e0fe0395", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280f8ef3-0f", "ovs_interfaceid": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.733329] env[62368]: DEBUG oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 768.733329] env[62368]: value = "vm-259765" [ 768.733329] env[62368]: _type = "VirtualMachine" [ 768.733329] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 768.733329] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1f52092f-a156-4b06-b08c-6dff4c632277 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.738021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-843d14d3-209b-477b-a91c-67b1dd78dc4e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.132s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.743927] env[62368]: DEBUG oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lease: (returnval){ [ 768.743927] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5255d60c-ed21-582a-9f72-7ffe98109c7f" [ 768.743927] env[62368]: _type = "HttpNfcLease" [ 768.743927] env[62368]: } obtained for exporting VM: (result){ [ 768.743927] env[62368]: value = "vm-259765" [ 768.743927] env[62368]: _type = "VirtualMachine" [ 768.743927] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 768.743927] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the lease: (returnval){ [ 768.743927] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5255d60c-ed21-582a-9f72-7ffe98109c7f" [ 768.743927] env[62368]: _type = "HttpNfcLease" [ 768.743927] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 768.753158] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 768.753158] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5255d60c-ed21-582a-9f72-7ffe98109c7f" [ 768.753158] env[62368]: _type = "HttpNfcLease" [ 768.753158] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 768.865413] env[62368]: DEBUG nova.network.neutron [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Successfully created port: 59b16e71-e7dd-438f-812a-c369af2f64be {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.036676] env[62368]: DEBUG nova.network.neutron [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.079323] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e8ab4a-85b6-4d7f-bba1-26569a2126ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.093161] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1cc1d7-c7b6-4e0d-921c-349b197aba79 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.131048] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6f0f86-7b1b-45d8-89ea-2ecea3da3d70 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.140216] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d36d7d3-c77f-409f-a498-0a3d8dca02b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.157104] env[62368]: DEBUG nova.compute.provider_tree [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.203555] env[62368]: DEBUG nova.network.neutron [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updating instance_info_cache with network_info: [{"id": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "address": "fa:16:3e:ba:65:a2", "network": {"id": "831eece9-e402-442b-b933-93e089c864be", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1971247571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd4cd004b0d54bb49a88722fd1312249", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3bbc3d-86", "ovs_interfaceid": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.226981] env[62368]: DEBUG oslo_concurrency.lockutils [req-30ff161d-8775-4548-b5f3-b76473fed80c req-31cab586-b23c-4e9c-a649-5f2b398b9b87 service nova] Releasing lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.240025] env[62368]: DEBUG nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.246920] env[62368]: DEBUG nova.compute.manager [req-38980836-2c70-4b52-a8a3-cc67a1d60ee5 req-7e56b44f-ecb3-4f90-896e-40ca4dea59df service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Received event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.246920] env[62368]: DEBUG nova.compute.manager [req-38980836-2c70-4b52-a8a3-cc67a1d60ee5 req-7e56b44f-ecb3-4f90-896e-40ca4dea59df service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing instance network info cache due to event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.246920] env[62368]: DEBUG oslo_concurrency.lockutils [req-38980836-2c70-4b52-a8a3-cc67a1d60ee5 req-7e56b44f-ecb3-4f90-896e-40ca4dea59df service nova] Acquiring lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.246920] env[62368]: DEBUG oslo_concurrency.lockutils [req-38980836-2c70-4b52-a8a3-cc67a1d60ee5 req-7e56b44f-ecb3-4f90-896e-40ca4dea59df service nova] Acquired lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.246920] env[62368]: DEBUG nova.network.neutron [req-38980836-2c70-4b52-a8a3-cc67a1d60ee5 req-7e56b44f-ecb3-4f90-896e-40ca4dea59df service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.257646] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 769.257646] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5255d60c-ed21-582a-9f72-7ffe98109c7f" [ 769.257646] env[62368]: _type = "HttpNfcLease" [ 769.257646] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 769.257941] env[62368]: DEBUG oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 769.257941] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5255d60c-ed21-582a-9f72-7ffe98109c7f" [ 769.257941] env[62368]: _type = "HttpNfcLease" [ 769.257941] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 769.259387] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d4ee1a-95f2-4e1c-a274-86844c5d9be7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.269452] env[62368]: DEBUG oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523b38b6-3ae1-80f5-9a19-8529046594f5/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 769.269672] env[62368]: DEBUG oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523b38b6-3ae1-80f5-9a19-8529046594f5/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 769.361554] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-333d610b-cc4a-4d3a-9a92-248e3da77117 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.436839] env[62368]: DEBUG nova.network.neutron [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Successfully created port: 64b32060-dc43-4d7e-88b2-84f2cb0e9055 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.663940] env[62368]: DEBUG nova.scheduler.client.report [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.707559] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Releasing lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.708098] env[62368]: DEBUG nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Instance network_info: |[{"id": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "address": "fa:16:3e:ba:65:a2", "network": {"id": "831eece9-e402-442b-b933-93e089c864be", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1971247571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd4cd004b0d54bb49a88722fd1312249", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3bbc3d-86", "ovs_interfaceid": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 769.708431] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] Acquired lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.708584] env[62368]: DEBUG nova.network.neutron [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Refreshing network info cache for port 4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.712913] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:65:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8e028024-a9c1-4cae-8849-ea770a7ae0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f3bbc3d-86d4-442f-9f71-dc40a130ce4c', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.731563] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Creating folder: Project (fd4cd004b0d54bb49a88722fd1312249). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 769.738794] env[62368]: DEBUG nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 769.749729] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28d42b84-51a4-450d-b9a9-8260fbfb2c94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.775418] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Created folder: Project (fd4cd004b0d54bb49a88722fd1312249) in parent group-v259706. [ 769.775418] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Creating folder: Instances. Parent ref: group-v259766. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 769.778337] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 769.778829] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 769.780108] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.780108] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 769.780108] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.780108] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 769.780108] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 769.780298] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 769.780324] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 769.780698] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 769.780698] env[62368]: DEBUG nova.virt.hardware [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.781142] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00f0a16e-deed-4351-822f-26a0035145f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.785007] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd006a0-6a71-4ced-8f14-66d926fddf55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.789033] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.796999] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12572781-ed0d-41cc-9b29-62d4a7aa9f03 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.806228] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Created folder: Instances in parent group-v259766. [ 769.806228] env[62368]: DEBUG oslo.service.loopingcall [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.806228] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 769.806317] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fb9b1ea-6f67-4c07-be77-635e24eec71a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.846731] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.846731] env[62368]: value = "task-1198335" [ 769.846731] env[62368]: _type = "Task" [ 769.846731] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.855626] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198335, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.928614] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.929023] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.929367] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.929765] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.930164] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.940410] env[62368]: INFO nova.compute.manager [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Terminating instance [ 769.945045] env[62368]: DEBUG nova.compute.manager [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 769.945045] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 769.946343] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d692f561-be44-4510-83a3-5ad043da1a39 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.959298] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 769.959826] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1115cf60-fb85-41b0-a3fb-4686e0e853d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.969263] env[62368]: DEBUG oslo_vmware.api [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 769.969263] env[62368]: value = "task-1198336" [ 769.969263] env[62368]: _type = "Task" [ 769.969263] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.979989] env[62368]: DEBUG oslo_vmware.api [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198336, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.173070] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.173070] env[62368]: DEBUG nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.174569] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.265s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.175360] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.181598] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.100s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.183672] env[62368]: INFO nova.compute.claims [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.218541] env[62368]: INFO nova.scheduler.client.report [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Deleted allocations for instance 4dc7b6b4-1bf5-4195-bb93-14756f8f9986 [ 770.360252] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198335, 'name': CreateVM_Task, 'duration_secs': 0.460099} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.360252] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 770.360252] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.360252] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.360252] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 770.360252] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54baabcd-c1a6-473d-aa1d-31ca926d9ed6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.366942] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 770.366942] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ce0d01-8d31-5c4c-63ac-da09acc7ca55" [ 770.366942] env[62368]: _type = "Task" [ 770.366942] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.372859] env[62368]: DEBUG nova.network.neutron [req-38980836-2c70-4b52-a8a3-cc67a1d60ee5 req-7e56b44f-ecb3-4f90-896e-40ca4dea59df service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updated VIF entry in instance network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.375644] env[62368]: DEBUG nova.network.neutron [req-38980836-2c70-4b52-a8a3-cc67a1d60ee5 req-7e56b44f-ecb3-4f90-896e-40ca4dea59df service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updating instance_info_cache with network_info: [{"id": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "address": "fa:16:3e:fc:46:9c", "network": {"id": "d71a42e9-1e1c-4abd-b585-2362ba1b8cfb", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1629701000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3b87e3b7d1d947b79f4834e4e0fe0395", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280f8ef3-0f", "ovs_interfaceid": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.390628] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ce0d01-8d31-5c4c-63ac-da09acc7ca55, 'name': SearchDatastore_Task, 'duration_secs': 0.010483} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.391595] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.391851] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.392327] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.392566] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.392789] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.393372] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88e672ad-c36c-43be-a1d4-58704d9210ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.404157] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.404157] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 770.404514] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-262ee20a-c71a-42ea-b4bc-e4cba47f0051 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.412108] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 770.412108] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52663953-1ec7-6342-8bb0-75d040f4f17f" [ 770.412108] env[62368]: _type = "Task" [ 770.412108] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.420784] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52663953-1ec7-6342-8bb0-75d040f4f17f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.482913] env[62368]: DEBUG oslo_vmware.api [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198336, 'name': PowerOffVM_Task, 'duration_secs': 0.234178} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.482913] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 770.482913] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 770.482913] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c1c7f5c-cdd1-4e31-b477-829f63003daa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.566962] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 770.567402] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 770.567752] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleting the datastore file [datastore1] af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.568167] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3fc0d00-ff3e-4387-b513-81fb3bbfbc81 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.576510] env[62368]: DEBUG oslo_vmware.api [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 770.576510] env[62368]: value = "task-1198338" [ 770.576510] env[62368]: _type = "Task" [ 770.576510] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.586143] env[62368]: DEBUG oslo_vmware.api [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.689087] env[62368]: DEBUG nova.compute.utils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.697027] env[62368]: DEBUG nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.697027] env[62368]: DEBUG nova.network.neutron [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 770.734287] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d7db579-85b9-4593-a63a-5e8d06773802 tempest-VolumesAdminNegativeTest-1206951801 tempest-VolumesAdminNegativeTest-1206951801-project-member] Lock "4dc7b6b4-1bf5-4195-bb93-14756f8f9986" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.945s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.877939] env[62368]: DEBUG oslo_concurrency.lockutils [req-38980836-2c70-4b52-a8a3-cc67a1d60ee5 req-7e56b44f-ecb3-4f90-896e-40ca4dea59df service nova] Releasing lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.929078] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52663953-1ec7-6342-8bb0-75d040f4f17f, 'name': SearchDatastore_Task, 'duration_secs': 0.00955} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.930299] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11d12978-f322-4a39-8457-63a5f5303b04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.939060] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 770.939060] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529ba0d8-7d8b-67e9-9e5c-42650e33776f" [ 770.939060] env[62368]: _type = "Task" [ 770.939060] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.949278] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529ba0d8-7d8b-67e9-9e5c-42650e33776f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.962159] env[62368]: DEBUG nova.network.neutron [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updated VIF entry in instance network info cache for port 4f3bbc3d-86d4-442f-9f71-dc40a130ce4c. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.962159] env[62368]: DEBUG nova.network.neutron [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updating instance_info_cache with network_info: [{"id": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "address": "fa:16:3e:ba:65:a2", "network": {"id": "831eece9-e402-442b-b933-93e089c864be", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1971247571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd4cd004b0d54bb49a88722fd1312249", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3bbc3d-86", "ovs_interfaceid": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.966741] env[62368]: DEBUG nova.policy [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0800ab273ca04fbf9396175b57eed6b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e821059910b4e32aab596c6f4d521d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.090780] env[62368]: DEBUG oslo_vmware.api [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150686} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.090780] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.090780] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 771.090780] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 771.090780] env[62368]: INFO nova.compute.manager [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Took 1.15 seconds to destroy the instance on the hypervisor. [ 771.091164] env[62368]: DEBUG oslo.service.loopingcall [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.091450] env[62368]: DEBUG nova.compute.manager [-] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.091569] env[62368]: DEBUG nova.network.neutron [-] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 771.197809] env[62368]: DEBUG nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 771.454324] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529ba0d8-7d8b-67e9-9e5c-42650e33776f, 'name': SearchDatastore_Task, 'duration_secs': 0.009835} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.454646] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.454914] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3/0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 771.455352] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a100230-4158-419e-9ae4-665eaf65d995 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.465158] env[62368]: DEBUG oslo_concurrency.lockutils [req-a9476297-a9ab-470d-aeb4-8ffde844746e req-b31c0ff8-24a3-4cb0-9beb-6ca2ba7bfd8c service nova] Releasing lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.467180] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 771.467180] env[62368]: value = "task-1198339" [ 771.467180] env[62368]: _type = "Task" [ 771.467180] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.477478] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.606096] env[62368]: DEBUG nova.compute.manager [req-bb48c03e-e3e7-474b-a250-242def8d4d72 req-2aa7ce44-aab8-431a-81c0-dbbc49b7c18b service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Received event network-vif-plugged-59b16e71-e7dd-438f-812a-c369af2f64be {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.606340] env[62368]: DEBUG oslo_concurrency.lockutils [req-bb48c03e-e3e7-474b-a250-242def8d4d72 req-2aa7ce44-aab8-431a-81c0-dbbc49b7c18b service nova] Acquiring lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.606559] env[62368]: DEBUG oslo_concurrency.lockutils [req-bb48c03e-e3e7-474b-a250-242def8d4d72 req-2aa7ce44-aab8-431a-81c0-dbbc49b7c18b service nova] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.606738] env[62368]: DEBUG oslo_concurrency.lockutils [req-bb48c03e-e3e7-474b-a250-242def8d4d72 req-2aa7ce44-aab8-431a-81c0-dbbc49b7c18b service nova] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.606911] env[62368]: DEBUG nova.compute.manager [req-bb48c03e-e3e7-474b-a250-242def8d4d72 req-2aa7ce44-aab8-431a-81c0-dbbc49b7c18b service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] No waiting events found dispatching network-vif-plugged-59b16e71-e7dd-438f-812a-c369af2f64be {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 771.607568] env[62368]: WARNING nova.compute.manager [req-bb48c03e-e3e7-474b-a250-242def8d4d72 req-2aa7ce44-aab8-431a-81c0-dbbc49b7c18b service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Received unexpected event network-vif-plugged-59b16e71-e7dd-438f-812a-c369af2f64be for instance with vm_state building and task_state spawning. [ 771.612462] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54021f6f-350b-49e6-83d1-bd8ea5610fb6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.623961] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5203ac-2b88-45a5-9852-fe5693deb4e6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.633790] env[62368]: DEBUG nova.compute.manager [req-b64059a6-eea4-40b7-a354-be11d7f34965 req-c0a1b5e1-8bd7-4d88-a33b-4a956b41be44 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Received event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.633997] env[62368]: DEBUG nova.compute.manager [req-b64059a6-eea4-40b7-a354-be11d7f34965 req-c0a1b5e1-8bd7-4d88-a33b-4a956b41be44 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing instance network info cache due to event network-changed-280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 771.634282] env[62368]: DEBUG oslo_concurrency.lockutils [req-b64059a6-eea4-40b7-a354-be11d7f34965 req-c0a1b5e1-8bd7-4d88-a33b-4a956b41be44 service nova] Acquiring lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.634436] env[62368]: DEBUG oslo_concurrency.lockutils [req-b64059a6-eea4-40b7-a354-be11d7f34965 req-c0a1b5e1-8bd7-4d88-a33b-4a956b41be44 service nova] Acquired lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.634693] env[62368]: DEBUG nova.network.neutron [req-b64059a6-eea4-40b7-a354-be11d7f34965 req-c0a1b5e1-8bd7-4d88-a33b-4a956b41be44 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Refreshing network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 771.644128] env[62368]: DEBUG nova.network.neutron [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Successfully updated port: 59b16e71-e7dd-438f-812a-c369af2f64be {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 771.679438] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119b0d0a-48a9-4f0c-b80b-5297a1d58a0a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.692419] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c48d64-c0c6-4143-94ca-00b5284a3f28 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.718975] env[62368]: DEBUG nova.compute.provider_tree [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.862091] env[62368]: DEBUG nova.network.neutron [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Successfully created port: c96365aa-f01b-493f-b516-a4e22b5b43e1 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 771.978736] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198339, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491065} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.979806] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3/0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 771.980012] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 771.986022] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3c2f696-f7cd-4667-966e-4e67c6659c7a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.991696] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 771.991696] env[62368]: value = "task-1198340" [ 771.991696] env[62368]: _type = "Task" [ 771.991696] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.004697] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198340, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.221794] env[62368]: DEBUG nova.scheduler.client.report [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.227023] env[62368]: DEBUG nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.261587] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.261966] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.262306] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.262792] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.262792] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.262954] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.263235] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.263418] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.263831] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.264029] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.264260] env[62368]: DEBUG nova.virt.hardware [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.265645] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3e8a82-f468-4c38-832a-fa9131e7ddd0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.274947] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1bf061-cccd-4456-818b-d052c125f48f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.419142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.419142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.419142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.419142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.419757] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.420417] env[62368]: INFO nova.compute.manager [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Terminating instance [ 772.423742] env[62368]: DEBUG nova.compute.manager [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 772.424173] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 772.427241] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df06e0c7-4b92-4e20-b1f8-9ebb1b68f088 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.438359] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 772.438359] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdcec47a-3387-4643-983c-7d9531f7a784 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.449791] env[62368]: DEBUG oslo_vmware.api [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 772.449791] env[62368]: value = "task-1198341" [ 772.449791] env[62368]: _type = "Task" [ 772.449791] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.461612] env[62368]: DEBUG oslo_vmware.api [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198341, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.486122] env[62368]: DEBUG nova.network.neutron [-] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.505327] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198340, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095387} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.505610] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.506498] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49e6e5c-e992-483b-81c4-45fcabaa46da {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.535582] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3/0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.536419] env[62368]: DEBUG nova.network.neutron [req-b64059a6-eea4-40b7-a354-be11d7f34965 req-c0a1b5e1-8bd7-4d88-a33b-4a956b41be44 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updated VIF entry in instance network info cache for port 280f8ef3-0f46-4846-9e44-fff59d473e8b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 772.536921] env[62368]: DEBUG nova.network.neutron [req-b64059a6-eea4-40b7-a354-be11d7f34965 req-c0a1b5e1-8bd7-4d88-a33b-4a956b41be44 service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updating instance_info_cache with network_info: [{"id": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "address": "fa:16:3e:fc:46:9c", "network": {"id": "d71a42e9-1e1c-4abd-b585-2362ba1b8cfb", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1629701000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3b87e3b7d1d947b79f4834e4e0fe0395", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f8442aa5-73db-4599-8564-b98a6ea26b9b", "external-id": "nsx-vlan-transportzone-893", "segmentation_id": 893, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280f8ef3-0f", "ovs_interfaceid": "280f8ef3-0f46-4846-9e44-fff59d473e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.538144] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11d15949-633b-4fb8-af3d-1d1686321a0b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.564456] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 772.564456] env[62368]: value = "task-1198342" [ 772.564456] env[62368]: _type = "Task" [ 772.564456] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.575349] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198342, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.731490] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.732069] env[62368]: DEBUG nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 772.737021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.608s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.737021] env[62368]: INFO nova.compute.claims [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.959998] env[62368]: DEBUG oslo_vmware.api [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198341, 'name': PowerOffVM_Task, 'duration_secs': 0.229633} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.960306] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 772.960478] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 772.960741] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d9205b0-346d-4a01-8aaa-911460593b4f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.990034] env[62368]: INFO nova.compute.manager [-] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Took 1.90 seconds to deallocate network for instance. [ 773.055464] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 773.055928] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 773.056041] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Deleting the datastore file [datastore2] 4fe7b1a2-f894-4131-9456-fb4df8a4532d {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.056395] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-991132ad-1e6c-4eec-b7ee-5f6a361f83e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.061532] env[62368]: DEBUG oslo_concurrency.lockutils [req-b64059a6-eea4-40b7-a354-be11d7f34965 req-c0a1b5e1-8bd7-4d88-a33b-4a956b41be44 service nova] Releasing lock "refresh_cache-4fe7b1a2-f894-4131-9456-fb4df8a4532d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.072011] env[62368]: DEBUG oslo_vmware.api [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for the task: (returnval){ [ 773.072011] env[62368]: value = "task-1198344" [ 773.072011] env[62368]: _type = "Task" [ 773.072011] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.080298] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198342, 'name': ReconfigVM_Task, 'duration_secs': 0.346227} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.083250] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3/0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.083250] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-878a35eb-d3fc-4603-9c3c-8883260ce317 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.091665] env[62368]: DEBUG oslo_vmware.api [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198344, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.099110] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 773.099110] env[62368]: value = "task-1198345" [ 773.099110] env[62368]: _type = "Task" [ 773.099110] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.112146] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198345, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.241882] env[62368]: DEBUG nova.compute.utils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.251060] env[62368]: DEBUG nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 773.251536] env[62368]: DEBUG nova.network.neutron [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 773.342985] env[62368]: DEBUG nova.policy [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68c2209271e043b698fd3794955f40bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcecae6c09124e9fa4dcd5d63833b0bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 773.502234] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.589434] env[62368]: DEBUG oslo_vmware.api [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Task: {'id': task-1198344, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196229} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.590443] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 773.590651] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 773.590905] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 773.591198] env[62368]: INFO nova.compute.manager [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 773.593417] env[62368]: DEBUG oslo.service.loopingcall [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.593417] env[62368]: DEBUG nova.compute.manager [-] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.593417] env[62368]: DEBUG nova.network.neutron [-] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 773.610570] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198345, 'name': Rename_Task, 'duration_secs': 0.201734} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.610935] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 773.611242] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7af1db57-befc-40ea-8fc2-97b12dc35f25 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.621660] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 773.621660] env[62368]: value = "task-1198346" [ 773.621660] env[62368]: _type = "Task" [ 773.621660] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.633058] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198346, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.750499] env[62368]: DEBUG nova.network.neutron [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Successfully updated port: 64b32060-dc43-4d7e-88b2-84f2cb0e9055 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 773.757494] env[62368]: DEBUG nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 773.847379] env[62368]: DEBUG nova.objects.instance [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lazy-loading 'flavor' on Instance uuid 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.874322] env[62368]: DEBUG nova.compute.manager [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Received event network-changed-59b16e71-e7dd-438f-812a-c369af2f64be {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.874597] env[62368]: DEBUG nova.compute.manager [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Refreshing instance network info cache due to event network-changed-59b16e71-e7dd-438f-812a-c369af2f64be. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.874876] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] Acquiring lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.875101] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] Acquired lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.875281] env[62368]: DEBUG nova.network.neutron [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Refreshing network info cache for port 59b16e71-e7dd-438f-812a-c369af2f64be {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.894252] env[62368]: DEBUG nova.compute.manager [req-26d62727-9589-4c2e-bbb5-22195d10220d req-704f5c7a-8ed3-4657-86c7-a9dd64331a07 service nova] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Received event network-vif-deleted-b829cb49-cde3-4f16-8e08-6f56dff38e09 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.923550] env[62368]: DEBUG nova.network.neutron [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Successfully updated port: c96365aa-f01b-493f-b516-a4e22b5b43e1 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 774.102943] env[62368]: DEBUG nova.compute.manager [req-576ac3c7-8ad4-4ce8-b890-9d73075da914 req-b5972a69-fa67-4728-923d-bb073bf814cb service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Received event network-vif-plugged-c96365aa-f01b-493f-b516-a4e22b5b43e1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.103217] env[62368]: DEBUG oslo_concurrency.lockutils [req-576ac3c7-8ad4-4ce8-b890-9d73075da914 req-b5972a69-fa67-4728-923d-bb073bf814cb service nova] Acquiring lock "3f1af54b-392f-432a-9ffa-a133da428f94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.103379] env[62368]: DEBUG oslo_concurrency.lockutils [req-576ac3c7-8ad4-4ce8-b890-9d73075da914 req-b5972a69-fa67-4728-923d-bb073bf814cb service nova] Lock "3f1af54b-392f-432a-9ffa-a133da428f94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.103595] env[62368]: DEBUG oslo_concurrency.lockutils [req-576ac3c7-8ad4-4ce8-b890-9d73075da914 req-b5972a69-fa67-4728-923d-bb073bf814cb service nova] Lock "3f1af54b-392f-432a-9ffa-a133da428f94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.103774] env[62368]: DEBUG nova.compute.manager [req-576ac3c7-8ad4-4ce8-b890-9d73075da914 req-b5972a69-fa67-4728-923d-bb073bf814cb service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] No waiting events found dispatching network-vif-plugged-c96365aa-f01b-493f-b516-a4e22b5b43e1 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 774.104219] env[62368]: WARNING nova.compute.manager [req-576ac3c7-8ad4-4ce8-b890-9d73075da914 req-b5972a69-fa67-4728-923d-bb073bf814cb service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Received unexpected event network-vif-plugged-c96365aa-f01b-493f-b516-a4e22b5b43e1 for instance with vm_state building and task_state spawning. [ 774.131543] env[62368]: DEBUG nova.network.neutron [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Successfully created port: 6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.144096] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198346, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.161665] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609067e4-4e49-43a3-b3c2-dba966a0ff15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.170912] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c901ad-a4ba-4cc6-bc9b-feb9e39ec727 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.222267] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf502f7-7c40-45e3-ab88-94e6506d4943 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.234873] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3122b6-10be-4bbf-88ad-6c2fa8082935 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.250553] env[62368]: DEBUG nova.compute.provider_tree [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.255356] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.355842] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.356124] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquired lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.428998] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-3f1af54b-392f-432a-9ffa-a133da428f94" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.428998] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-3f1af54b-392f-432a-9ffa-a133da428f94" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.428998] env[62368]: DEBUG nova.network.neutron [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.487273] env[62368]: DEBUG nova.network.neutron [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.639455] env[62368]: DEBUG oslo_vmware.api [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198346, 'name': PowerOnVM_Task, 'duration_secs': 0.536881} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.640054] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 774.640400] env[62368]: INFO nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Took 7.38 seconds to spawn the instance on the hypervisor. [ 774.640953] env[62368]: DEBUG nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 774.641934] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217e3a64-b251-436a-a644-7fec46c4e57f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.759016] env[62368]: DEBUG nova.scheduler.client.report [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.761774] env[62368]: DEBUG nova.network.neutron [-] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.774701] env[62368]: DEBUG nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 774.808822] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.808822] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.808822] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.809038] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.809283] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.809613] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.809979] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.810407] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.810711] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.810997] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.811299] env[62368]: DEBUG nova.virt.hardware [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.814105] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12afdb45-6504-4e2f-a39c-de6e1588915c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.822994] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7a3b7d-fe06-453c-a044-9bf3ac2517e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.881681] env[62368]: DEBUG nova.network.neutron [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.166021] env[62368]: INFO nova.compute.manager [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Took 36.40 seconds to build instance. [ 775.218339] env[62368]: DEBUG nova.network.neutron [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.265060] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.265060] env[62368]: DEBUG nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 775.269513] env[62368]: INFO nova.compute.manager [-] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Took 1.68 seconds to deallocate network for instance. [ 775.269513] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.086s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.272149] env[62368]: INFO nova.compute.claims [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.385094] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] Releasing lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.385704] env[62368]: DEBUG nova.compute.manager [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Received event network-vif-plugged-64b32060-dc43-4d7e-88b2-84f2cb0e9055 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.386034] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] Acquiring lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.386426] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.386643] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.386848] env[62368]: DEBUG nova.compute.manager [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] No waiting events found dispatching network-vif-plugged-64b32060-dc43-4d7e-88b2-84f2cb0e9055 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 775.387037] env[62368]: WARNING nova.compute.manager [req-9f820e8b-53ad-48e1-91ef-698c810659ec req-06f9d422-8ee8-4812-8bef-e838470bd1ea service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Received unexpected event network-vif-plugged-64b32060-dc43-4d7e-88b2-84f2cb0e9055 for instance with vm_state building and task_state spawning. [ 775.387426] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquired lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.387581] env[62368]: DEBUG nova.network.neutron [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.511853] env[62368]: DEBUG nova.network.neutron [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Updating instance_info_cache with network_info: [{"id": "c96365aa-f01b-493f-b516-a4e22b5b43e1", "address": "fa:16:3e:2a:ad:8e", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc96365aa-f0", "ovs_interfaceid": "c96365aa-f01b-493f-b516-a4e22b5b43e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.576368] env[62368]: DEBUG nova.network.neutron [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.668226] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e04c3b0b-7aeb-487a-85ad-857e39fb42b8 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.650s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.776522] env[62368]: DEBUG nova.compute.utils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.784165] env[62368]: DEBUG nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 775.784165] env[62368]: DEBUG nova.network.neutron [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 775.785695] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.831563] env[62368]: DEBUG nova.policy [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6aa8db505789477ca7756d24a1636e2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f00de1923334a2999f52d0a374cd522', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.953105] env[62368]: DEBUG nova.network.neutron [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.011360] env[62368]: DEBUG nova.compute.manager [req-6ddc32b0-af91-45db-a4fe-3eab3080b672 req-8c543cc4-3d6e-410f-80f7-39246e6e4388 service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Received event network-changed-64b32060-dc43-4d7e-88b2-84f2cb0e9055 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.011693] env[62368]: DEBUG nova.compute.manager [req-6ddc32b0-af91-45db-a4fe-3eab3080b672 req-8c543cc4-3d6e-410f-80f7-39246e6e4388 service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Refreshing instance network info cache due to event network-changed-64b32060-dc43-4d7e-88b2-84f2cb0e9055. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.011986] env[62368]: DEBUG oslo_concurrency.lockutils [req-6ddc32b0-af91-45db-a4fe-3eab3080b672 req-8c543cc4-3d6e-410f-80f7-39246e6e4388 service nova] Acquiring lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.014166] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-3f1af54b-392f-432a-9ffa-a133da428f94" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.014468] env[62368]: DEBUG nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Instance network_info: |[{"id": "c96365aa-f01b-493f-b516-a4e22b5b43e1", "address": "fa:16:3e:2a:ad:8e", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc96365aa-f0", "ovs_interfaceid": "c96365aa-f01b-493f-b516-a4e22b5b43e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 776.014841] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:ad:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c96365aa-f01b-493f-b516-a4e22b5b43e1', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.023987] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating folder: Project (7e821059910b4e32aab596c6f4d521d8). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 776.027129] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3576cd7b-22dd-4cf1-b693-c420b970e42a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.042195] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created folder: Project (7e821059910b4e32aab596c6f4d521d8) in parent group-v259706. [ 776.042583] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating folder: Instances. Parent ref: group-v259769. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 776.043284] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5f5458e-508d-4a47-972f-175b0e1d9099 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.056462] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created folder: Instances in parent group-v259769. [ 776.056787] env[62368]: DEBUG oslo.service.loopingcall [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.057030] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 776.057295] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c7bf2a1-f5b4-48c1-b7b4-a8cf4e91e66f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.087865] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.087865] env[62368]: value = "task-1198349" [ 776.087865] env[62368]: _type = "Task" [ 776.087865] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.100441] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198349, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.178853] env[62368]: DEBUG nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.293085] env[62368]: DEBUG nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 776.372176] env[62368]: DEBUG nova.compute.manager [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Received event network-changed-c96365aa-f01b-493f-b516-a4e22b5b43e1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.372581] env[62368]: DEBUG nova.compute.manager [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Refreshing instance network info cache due to event network-changed-c96365aa-f01b-493f-b516-a4e22b5b43e1. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.372751] env[62368]: DEBUG oslo_concurrency.lockutils [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] Acquiring lock "refresh_cache-3f1af54b-392f-432a-9ffa-a133da428f94" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.374064] env[62368]: DEBUG oslo_concurrency.lockutils [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] Acquired lock "refresh_cache-3f1af54b-392f-432a-9ffa-a133da428f94" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.374064] env[62368]: DEBUG nova.network.neutron [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Refreshing network info cache for port c96365aa-f01b-493f-b516-a4e22b5b43e1 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 776.601344] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198349, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.607798] env[62368]: DEBUG nova.network.neutron [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Successfully created port: e7892749-2fef-4fa0-ba5b-348c58477c93 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.670986] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148e7bf3-7c87-4315-a7d8-c37e7e9f1795 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.681906] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e377dc-ac37-433d-b7a7-b968d4b8e95d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.689205] env[62368]: DEBUG nova.network.neutron [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Successfully updated port: 6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.728833] env[62368]: DEBUG nova.network.neutron [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.731120] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.731936] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c163f663-e30d-46ad-b8da-692a6fad4c80 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.743359] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368c0db8-86b7-4716-8340-73ea79771bc9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.748842] env[62368]: DEBUG nova.network.neutron [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Updating instance_info_cache with network_info: [{"id": "59b16e71-e7dd-438f-812a-c369af2f64be", "address": "fa:16:3e:84:e1:3d", "network": {"id": "b427bc97-b4a3-4b87-b62f-56050326e00b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-808856501", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.224", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ff43f079c154fc6982d0cd583c6f789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59b16e71-e7", "ovs_interfaceid": "59b16e71-e7dd-438f-812a-c369af2f64be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "64b32060-dc43-4d7e-88b2-84f2cb0e9055", "address": "fa:16:3e:66:44:94", "network": {"id": "887ae77e-3b75-4e2b-b2f9-0f55e0b191b6", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-915809281", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "0ff43f079c154fc6982d0cd583c6f789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64b32060-dc", "ovs_interfaceid": "64b32060-dc43-4d7e-88b2-84f2cb0e9055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.764849] env[62368]: DEBUG nova.compute.provider_tree [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.815831] env[62368]: DEBUG nova.compute.manager [req-6cef054b-88ea-4963-832b-00305ac770a4 req-3fa2fa06-0525-44da-9d72-b8d49b1ab850 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-vif-plugged-6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.816127] env[62368]: DEBUG oslo_concurrency.lockutils [req-6cef054b-88ea-4963-832b-00305ac770a4 req-3fa2fa06-0525-44da-9d72-b8d49b1ab850 service nova] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.816661] env[62368]: DEBUG oslo_concurrency.lockutils [req-6cef054b-88ea-4963-832b-00305ac770a4 req-3fa2fa06-0525-44da-9d72-b8d49b1ab850 service nova] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.816661] env[62368]: DEBUG oslo_concurrency.lockutils [req-6cef054b-88ea-4963-832b-00305ac770a4 req-3fa2fa06-0525-44da-9d72-b8d49b1ab850 service nova] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.816782] env[62368]: DEBUG nova.compute.manager [req-6cef054b-88ea-4963-832b-00305ac770a4 req-3fa2fa06-0525-44da-9d72-b8d49b1ab850 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] No waiting events found dispatching network-vif-plugged-6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 776.817009] env[62368]: WARNING nova.compute.manager [req-6cef054b-88ea-4963-832b-00305ac770a4 req-3fa2fa06-0525-44da-9d72-b8d49b1ab850 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received unexpected event network-vif-plugged-6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 for instance with vm_state building and task_state spawning. [ 777.107303] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198349, 'name': CreateVM_Task, 'duration_secs': 0.518992} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.107873] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 777.108752] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.108796] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.111646] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 777.111646] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64df8cd1-fa67-48b6-9bbb-22d8c7ff4441 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.122017] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 777.122017] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524f91a9-14aa-c85c-dde2-4f62baa21af6" [ 777.122017] env[62368]: _type = "Task" [ 777.122017] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.134118] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524f91a9-14aa-c85c-dde2-4f62baa21af6, 'name': SearchDatastore_Task, 'duration_secs': 0.011208} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.134805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.134886] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 777.135632] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.135632] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.135632] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 777.136070] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5060aa8e-275e-45af-b848-7004bb1b7345 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.146929] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 777.147149] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 777.147932] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a828a44f-4325-48dd-b18f-7da781f3d73b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.155711] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 777.155711] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5267b89d-d5f6-068b-401a-b13eb7c6fd70" [ 777.155711] env[62368]: _type = "Task" [ 777.155711] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.165426] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5267b89d-d5f6-068b-401a-b13eb7c6fd70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.194655] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.194897] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.194966] env[62368]: DEBUG nova.network.neutron [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 777.238655] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Releasing lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.238655] env[62368]: DEBUG nova.compute.manager [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Inject network info {{(pid=62368) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 777.239215] env[62368]: DEBUG nova.compute.manager [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] network_info to inject: |[{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 777.244876] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Reconfiguring VM instance to set the machine id {{(pid=62368) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 777.248683] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1d72040-3c26-462c-b2a0-8042c5fbcf29 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.260600] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Releasing lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.260948] env[62368]: DEBUG nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Instance network_info: |[{"id": "59b16e71-e7dd-438f-812a-c369af2f64be", "address": "fa:16:3e:84:e1:3d", "network": {"id": "b427bc97-b4a3-4b87-b62f-56050326e00b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-808856501", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.224", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ff43f079c154fc6982d0cd583c6f789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59b16e71-e7", "ovs_interfaceid": "59b16e71-e7dd-438f-812a-c369af2f64be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "64b32060-dc43-4d7e-88b2-84f2cb0e9055", "address": "fa:16:3e:66:44:94", "network": {"id": "887ae77e-3b75-4e2b-b2f9-0f55e0b191b6", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-915809281", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "0ff43f079c154fc6982d0cd583c6f789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64b32060-dc", "ovs_interfaceid": "64b32060-dc43-4d7e-88b2-84f2cb0e9055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 777.262476] env[62368]: DEBUG oslo_concurrency.lockutils [req-6ddc32b0-af91-45db-a4fe-3eab3080b672 req-8c543cc4-3d6e-410f-80f7-39246e6e4388 service nova] Acquired lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.262476] env[62368]: DEBUG nova.network.neutron [req-6ddc32b0-af91-45db-a4fe-3eab3080b672 req-8c543cc4-3d6e-410f-80f7-39246e6e4388 service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Refreshing network info cache for port 64b32060-dc43-4d7e-88b2-84f2cb0e9055 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 777.263265] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:e1:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00a15667-7ca5-4dc9-be92-164750d87988', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59b16e71-e7dd-438f-812a-c369af2f64be', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:44:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50886eea-591a-452c-a27b-5f22cfc9df85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '64b32060-dc43-4d7e-88b2-84f2cb0e9055', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.273970] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Creating folder: Project (0ff43f079c154fc6982d0cd583c6f789). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 777.275928] env[62368]: DEBUG nova.scheduler.client.report [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.279329] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46bc386e-19ad-478c-a9b6-a8545b803e9b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.283279] env[62368]: DEBUG oslo_vmware.api [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 777.283279] env[62368]: value = "task-1198350" [ 777.283279] env[62368]: _type = "Task" [ 777.283279] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.297780] env[62368]: DEBUG oslo_vmware.api [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198350, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.299155] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Created folder: Project (0ff43f079c154fc6982d0cd583c6f789) in parent group-v259706. [ 777.300029] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Creating folder: Instances. Parent ref: group-v259772. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 777.300274] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b002a0c-8d02-4c59-a5ab-05c558601a3b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.303155] env[62368]: DEBUG nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 777.318988] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Created folder: Instances in parent group-v259772. [ 777.319325] env[62368]: DEBUG oslo.service.loopingcall [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.319514] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 777.319711] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-302ff0a6-0a55-4936-ba70-ffcfcc541361 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.341764] env[62368]: DEBUG nova.network.neutron [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Updated VIF entry in instance network info cache for port c96365aa-f01b-493f-b516-a4e22b5b43e1. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 777.342122] env[62368]: DEBUG nova.network.neutron [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Updating instance_info_cache with network_info: [{"id": "c96365aa-f01b-493f-b516-a4e22b5b43e1", "address": "fa:16:3e:2a:ad:8e", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc96365aa-f0", "ovs_interfaceid": "c96365aa-f01b-493f-b516-a4e22b5b43e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.350915] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.350915] env[62368]: value = "task-1198353" [ 777.350915] env[62368]: _type = "Task" [ 777.350915] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.357026] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.357282] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.358472] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.358472] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.358472] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.358472] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.358472] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.358472] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.358472] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.360319] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.360319] env[62368]: DEBUG nova.virt.hardware [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.360319] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d225ab76-5a41-4243-84b4-acba2db7b14d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.373312] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198353, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.374697] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab74220-e642-46e4-8c14-9ef22c580d31 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.402613] env[62368]: DEBUG oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523b38b6-3ae1-80f5-9a19-8529046594f5/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 777.403518] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82eafa6a-7072-4dec-b198-f1d7afc7d30a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.410742] env[62368]: DEBUG oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523b38b6-3ae1-80f5-9a19-8529046594f5/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 777.410951] env[62368]: ERROR oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523b38b6-3ae1-80f5-9a19-8529046594f5/disk-0.vmdk due to incomplete transfer. [ 777.411300] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-68f62389-ecb4-46bf-912f-638b0c081f2f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.419881] env[62368]: DEBUG oslo_vmware.rw_handles [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523b38b6-3ae1-80f5-9a19-8529046594f5/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 777.420253] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Uploaded image 6281ea63-c477-4e65-8fac-7b43765b0423 to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 777.423319] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 777.423748] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6214ba45-9ce4-4b46-afec-c6db55d5a1fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.432840] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 777.432840] env[62368]: value = "task-1198354" [ 777.432840] env[62368]: _type = "Task" [ 777.432840] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.442341] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198354, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.523258] env[62368]: DEBUG nova.objects.instance [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lazy-loading 'flavor' on Instance uuid 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 777.668237] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5267b89d-d5f6-068b-401a-b13eb7c6fd70, 'name': SearchDatastore_Task, 'duration_secs': 0.010099} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.669563] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-548af236-f04b-4225-9241-1e49be9bf6a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.676399] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 777.676399] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5277f5df-7015-5829-fd99-8a1b140394de" [ 777.676399] env[62368]: _type = "Task" [ 777.676399] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.684977] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5277f5df-7015-5829-fd99-8a1b140394de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.745309] env[62368]: DEBUG nova.network.neutron [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.782589] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.783126] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.786168] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.714s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.786380] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.789176] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.663s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.789406] env[62368]: DEBUG nova.objects.instance [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lazy-loading 'resources' on Instance uuid 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 777.802733] env[62368]: DEBUG oslo_vmware.api [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198350, 'name': ReconfigVM_Task, 'duration_secs': 0.202167} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.807785] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7f9e3f0f-75ec-4e86-9b47-6f9b9bc6d677 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Reconfigured VM instance to set the machine id {{(pid=62368) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 777.816109] env[62368]: INFO nova.scheduler.client.report [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleted allocations for instance e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0 [ 777.845220] env[62368]: DEBUG oslo_concurrency.lockutils [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] Releasing lock "refresh_cache-3f1af54b-392f-432a-9ffa-a133da428f94" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.846275] env[62368]: DEBUG nova.compute.manager [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Received event network-vif-deleted-280f8ef3-0f46-4846-9e44-fff59d473e8b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.846275] env[62368]: DEBUG nova.compute.manager [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Received event network-changed-d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.846275] env[62368]: DEBUG nova.compute.manager [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Refreshing instance network info cache due to event network-changed-d98e888b-22dd-483e-999f-4e91c0e15ab6. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 777.846275] env[62368]: DEBUG oslo_concurrency.lockutils [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] Acquiring lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.846511] env[62368]: DEBUG oslo_concurrency.lockutils [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] Acquired lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.847517] env[62368]: DEBUG nova.network.neutron [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Refreshing network info cache for port d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 777.864382] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198353, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.946804] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198354, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.994828] env[62368]: DEBUG nova.network.neutron [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.028984] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.039938] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.040794] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.041058] env[62368]: INFO nova.compute.manager [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Rebooting instance [ 778.043292] env[62368]: DEBUG nova.compute.manager [req-afe34d98-3580-42b4-a6f6-23e66c2945d4 req-524c10e9-6fef-4b09-baa5-6161723f80ab service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Received event network-changed-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.043530] env[62368]: DEBUG nova.compute.manager [req-afe34d98-3580-42b4-a6f6-23e66c2945d4 req-524c10e9-6fef-4b09-baa5-6161723f80ab service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Refreshing instance network info cache due to event network-changed-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 778.044227] env[62368]: DEBUG oslo_concurrency.lockutils [req-afe34d98-3580-42b4-a6f6-23e66c2945d4 req-524c10e9-6fef-4b09-baa5-6161723f80ab service nova] Acquiring lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.044227] env[62368]: DEBUG oslo_concurrency.lockutils [req-afe34d98-3580-42b4-a6f6-23e66c2945d4 req-524c10e9-6fef-4b09-baa5-6161723f80ab service nova] Acquired lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.044227] env[62368]: DEBUG nova.network.neutron [req-afe34d98-3580-42b4-a6f6-23e66c2945d4 req-524c10e9-6fef-4b09-baa5-6161723f80ab service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Refreshing network info cache for port 4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.102488] env[62368]: DEBUG nova.network.neutron [req-6ddc32b0-af91-45db-a4fe-3eab3080b672 req-8c543cc4-3d6e-410f-80f7-39246e6e4388 service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Updated VIF entry in instance network info cache for port 64b32060-dc43-4d7e-88b2-84f2cb0e9055. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 778.102920] env[62368]: DEBUG nova.network.neutron [req-6ddc32b0-af91-45db-a4fe-3eab3080b672 req-8c543cc4-3d6e-410f-80f7-39246e6e4388 service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Updating instance_info_cache with network_info: [{"id": "59b16e71-e7dd-438f-812a-c369af2f64be", "address": "fa:16:3e:84:e1:3d", "network": {"id": "b427bc97-b4a3-4b87-b62f-56050326e00b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-808856501", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.224", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ff43f079c154fc6982d0cd583c6f789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59b16e71-e7", "ovs_interfaceid": "59b16e71-e7dd-438f-812a-c369af2f64be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "64b32060-dc43-4d7e-88b2-84f2cb0e9055", "address": "fa:16:3e:66:44:94", "network": {"id": "887ae77e-3b75-4e2b-b2f9-0f55e0b191b6", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-915809281", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "0ff43f079c154fc6982d0cd583c6f789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64b32060-dc", "ovs_interfaceid": "64b32060-dc43-4d7e-88b2-84f2cb0e9055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.193264] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5277f5df-7015-5829-fd99-8a1b140394de, 'name': SearchDatastore_Task, 'duration_secs': 0.031793} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.193264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.193264] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 3f1af54b-392f-432a-9ffa-a133da428f94/3f1af54b-392f-432a-9ffa-a133da428f94.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 778.193264] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-860c4021-4e0b-4d36-8707-b1947af35628 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.203862] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 778.203862] env[62368]: value = "task-1198355" [ 778.203862] env[62368]: _type = "Task" [ 778.203862] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.216937] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198355, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.293583] env[62368]: DEBUG nova.compute.utils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.300420] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 778.300420] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 778.330601] env[62368]: DEBUG oslo_concurrency.lockutils [None req-491ba54c-f91d-4353-ac80-31abcf039a86 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.236s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.370732] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198353, 'name': CreateVM_Task, 'duration_secs': 0.523581} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.370856] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 778.371718] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.371841] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.372164] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 778.372424] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-260304e6-9694-4ec7-829a-1ec27a232838 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.386745] env[62368]: DEBUG nova.policy [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e17ec74ebf8d4c12accd841e4503b0df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23e257a2707d4c5cb3062081be3c0546', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 778.388404] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 778.388404] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e1d7cb-87d5-fb5c-8ace-a554621449a1" [ 778.388404] env[62368]: _type = "Task" [ 778.388404] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.400484] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e1d7cb-87d5-fb5c-8ace-a554621449a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.450036] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198354, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.500228] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.500687] env[62368]: DEBUG nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Instance network_info: |[{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 778.505709] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:4e:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a1f27ab-91ba-4185-9fdb-c5e31d622ed6', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.516707] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Creating folder: Project (fcecae6c09124e9fa4dcd5d63833b0bd). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 778.520053] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-621db64a-3c22-47aa-8210-7cb37d947ab1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.536534] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Created folder: Project (fcecae6c09124e9fa4dcd5d63833b0bd) in parent group-v259706. [ 778.536534] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Creating folder: Instances. Parent ref: group-v259775. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 778.536534] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbf99096-dd15-4f47-b569-8526dc42a92e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.556772] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Created folder: Instances in parent group-v259775. [ 778.556827] env[62368]: DEBUG oslo.service.loopingcall [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.557382] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 778.557599] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d619cd9b-ba1c-420a-a1b2-b50538f0d05d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.585464] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.595022] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.595022] env[62368]: value = "task-1198358" [ 778.595022] env[62368]: _type = "Task" [ 778.595022] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.605633] env[62368]: DEBUG oslo_concurrency.lockutils [req-6ddc32b0-af91-45db-a4fe-3eab3080b672 req-8c543cc4-3d6e-410f-80f7-39246e6e4388 service nova] Releasing lock "refresh_cache-300c6c32-b365-40c2-bc7d-6ce028f10476" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.606391] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198358, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.676388] env[62368]: DEBUG nova.compute.manager [req-8bac9d81-8020-456e-b63a-8a4b5ac3e640 req-3a6777d3-ff60-46be-b34f-30f805e3a168 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-changed-6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.676867] env[62368]: DEBUG nova.compute.manager [req-8bac9d81-8020-456e-b63a-8a4b5ac3e640 req-3a6777d3-ff60-46be-b34f-30f805e3a168 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Refreshing instance network info cache due to event network-changed-6a1f27ab-91ba-4185-9fdb-c5e31d622ed6. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 778.677306] env[62368]: DEBUG oslo_concurrency.lockutils [req-8bac9d81-8020-456e-b63a-8a4b5ac3e640 req-3a6777d3-ff60-46be-b34f-30f805e3a168 service nova] Acquiring lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.677566] env[62368]: DEBUG oslo_concurrency.lockutils [req-8bac9d81-8020-456e-b63a-8a4b5ac3e640 req-3a6777d3-ff60-46be-b34f-30f805e3a168 service nova] Acquired lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.677871] env[62368]: DEBUG nova.network.neutron [req-8bac9d81-8020-456e-b63a-8a4b5ac3e640 req-3a6777d3-ff60-46be-b34f-30f805e3a168 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Refreshing network info cache for port 6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.718666] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198355, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.736358] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bc6dd8-effb-47c9-8ebd-da8fd91494e6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.748764] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1466b8ba-6898-4f16-95a1-aceb6c680e04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.795599] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef5dbdc-53ed-4e32-9ab2-89d273bfa659 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.798953] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.812274] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6ab1ca-6f40-4734-8394-dfba2129cfdf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.836386] env[62368]: DEBUG nova.compute.provider_tree [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.887175] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Successfully created port: 9897441b-2e19-4f96-8124-6aae0b31e308 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.905843] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e1d7cb-87d5-fb5c-8ace-a554621449a1, 'name': SearchDatastore_Task, 'duration_secs': 0.040448} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.906323] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.906641] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.906920] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.907109] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.907334] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.907709] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b89c6b5-18d0-4d12-b602-377501674d5c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.925658] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.925926] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 778.931828] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9677bd2-5713-4357-8b69-334baf2f8d8d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.942421] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 778.942421] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529ed78f-8c2e-8ae5-3f51-99036d029370" [ 778.942421] env[62368]: _type = "Task" [ 778.942421] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.950336] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198354, 'name': Destroy_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.956909] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529ed78f-8c2e-8ae5-3f51-99036d029370, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.110781] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198358, 'name': CreateVM_Task, 'duration_secs': 0.483556} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.110781] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 779.110897] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.111217] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.111641] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.112427] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b52834e-4b2f-4ee7-9918-ce18ffb0bb25 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.116330] env[62368]: DEBUG nova.network.neutron [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Successfully updated port: e7892749-2fef-4fa0-ba5b-348c58477c93 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 779.121092] env[62368]: DEBUG nova.network.neutron [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updated VIF entry in instance network info cache for port d98e888b-22dd-483e-999f-4e91c0e15ab6. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.121092] env[62368]: DEBUG nova.network.neutron [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.127631] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 779.127631] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521464cd-e44c-a75e-1592-06b9791841a7" [ 779.127631] env[62368]: _type = "Task" [ 779.127631] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.143341] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521464cd-e44c-a75e-1592-06b9791841a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.217881] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198355, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.85207} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.218237] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 3f1af54b-392f-432a-9ffa-a133da428f94/3f1af54b-392f-432a-9ffa-a133da428f94.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 779.218428] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 779.218693] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a287bda5-af76-43ed-a5cb-aac79fb13697 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.231196] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 779.231196] env[62368]: value = "task-1198359" [ 779.231196] env[62368]: _type = "Task" [ 779.231196] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.240725] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198359, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.286806] env[62368]: DEBUG nova.network.neutron [req-afe34d98-3580-42b4-a6f6-23e66c2945d4 req-524c10e9-6fef-4b09-baa5-6161723f80ab service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updated VIF entry in instance network info cache for port 4f3bbc3d-86d4-442f-9f71-dc40a130ce4c. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.287200] env[62368]: DEBUG nova.network.neutron [req-afe34d98-3580-42b4-a6f6-23e66c2945d4 req-524c10e9-6fef-4b09-baa5-6161723f80ab service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updating instance_info_cache with network_info: [{"id": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "address": "fa:16:3e:ba:65:a2", "network": {"id": "831eece9-e402-442b-b933-93e089c864be", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1971247571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd4cd004b0d54bb49a88722fd1312249", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3bbc3d-86", "ovs_interfaceid": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.311921] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.312173] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.340065] env[62368]: DEBUG nova.scheduler.client.report [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.452203] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198354, 'name': Destroy_Task, 'duration_secs': 1.552805} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.453233] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Destroyed the VM [ 779.453343] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 779.453670] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-aecf9bc3-26b9-4cb1-a795-4999765b132a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.459051] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529ed78f-8c2e-8ae5-3f51-99036d029370, 'name': SearchDatastore_Task, 'duration_secs': 0.055645} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.460226] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5204a05a-574c-4776-a911-472315ac39f8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.465160] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 779.465160] env[62368]: value = "task-1198360" [ 779.465160] env[62368]: _type = "Task" [ 779.465160] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.470066] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 779.470066] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5258c8bf-0cd9-876d-ff2e-0f87f945f8db" [ 779.470066] env[62368]: _type = "Task" [ 779.470066] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.477499] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198360, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.487058] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5258c8bf-0cd9-876d-ff2e-0f87f945f8db, 'name': SearchDatastore_Task, 'duration_secs': 0.011711} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.487441] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.487823] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 300c6c32-b365-40c2-bc7d-6ce028f10476/300c6c32-b365-40c2-bc7d-6ce028f10476.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 779.488183] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26e447c3-9982-455e-9a7c-af211a52d0f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.496981] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 779.496981] env[62368]: value = "task-1198361" [ 779.496981] env[62368]: _type = "Task" [ 779.496981] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.506298] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198361, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.525314] env[62368]: DEBUG nova.network.neutron [req-8bac9d81-8020-456e-b63a-8a4b5ac3e640 req-3a6777d3-ff60-46be-b34f-30f805e3a168 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updated VIF entry in instance network info cache for port 6a1f27ab-91ba-4185-9fdb-c5e31d622ed6. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.525695] env[62368]: DEBUG nova.network.neutron [req-8bac9d81-8020-456e-b63a-8a4b5ac3e640 req-3a6777d3-ff60-46be-b34f-30f805e3a168 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.619380] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "refresh_cache-0f1bff84-1fb9-471a-b685-7d527b376b0a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.619722] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquired lock "refresh_cache-0f1bff84-1fb9-471a-b685-7d527b376b0a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.619722] env[62368]: DEBUG nova.network.neutron [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.625398] env[62368]: DEBUG oslo_concurrency.lockutils [req-e384bf72-bf9a-4df0-aa29-4e9b664e47bd req-2296518e-dc29-4c3f-85b0-5ac1172639fa service nova] Releasing lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.625398] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquired lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.639962] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521464cd-e44c-a75e-1592-06b9791841a7, 'name': SearchDatastore_Task, 'duration_secs': 0.040745} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.640314] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.640591] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.640860] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.641124] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.641348] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.641659] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9149b7e-2e6f-4e2e-8099-ae02d97be556 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.652354] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.652354] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 779.653289] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2916bdb4-9a20-4fe1-96f3-71296d43fc98 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.660046] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 779.660046] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]523e8f1f-4181-84a3-db2c-e2724ee43349" [ 779.660046] env[62368]: _type = "Task" [ 779.660046] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.670226] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523e8f1f-4181-84a3-db2c-e2724ee43349, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.744606] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198359, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.229689} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.745405] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 779.746595] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f0dc8f-d9af-4884-95ba-7b4d24f1ce58 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.790365] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 3f1af54b-392f-432a-9ffa-a133da428f94/3f1af54b-392f-432a-9ffa-a133da428f94.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 779.791062] env[62368]: DEBUG oslo_concurrency.lockutils [req-afe34d98-3580-42b4-a6f6-23e66c2945d4 req-524c10e9-6fef-4b09-baa5-6161723f80ab service nova] Releasing lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.791502] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2fd8b2d-37c7-4720-a31e-38ac5d0adae4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.813452] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquired lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.813704] env[62368]: DEBUG nova.network.neutron [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.817921] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.827419] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 779.827419] env[62368]: value = "task-1198362" [ 779.827419] env[62368]: _type = "Task" [ 779.827419] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.846275] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198362, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.846275] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.057s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.854155] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.089s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.854505] env[62368]: DEBUG nova.objects.instance [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lazy-loading 'resources' on Instance uuid b393f654-0611-493e-a89f-445dfd3f8c13 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.869054] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.869054] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.869295] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.869351] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.869516] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.869774] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.870413] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.870551] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.870807] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.871015] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.871209] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.872727] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26434e8e-8b46-4d3e-8932-10c57c90fa39 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.886056] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdcc65e8-eca6-4401-bc3f-6c5035a19de9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.904570] env[62368]: INFO nova.scheduler.client.report [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Deleted allocations for instance 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4 [ 779.977986] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198360, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.008096] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198361, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.028677] env[62368]: DEBUG oslo_concurrency.lockutils [req-8bac9d81-8020-456e-b63a-8a4b5ac3e640 req-3a6777d3-ff60-46be-b34f-30f805e3a168 service nova] Releasing lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.172068] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523e8f1f-4181-84a3-db2c-e2724ee43349, 'name': SearchDatastore_Task, 'duration_secs': 0.068932} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.173544] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02eb75b4-12f7-47d9-b802-fb0490567589 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.182468] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 780.182468] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5215afea-0c35-ea90-7adf-e5cacb4ee6dc" [ 780.182468] env[62368]: _type = "Task" [ 780.182468] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.186343] env[62368]: DEBUG nova.network.neutron [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.205022] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5215afea-0c35-ea90-7adf-e5cacb4ee6dc, 'name': SearchDatastore_Task, 'duration_secs': 0.010871} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.205022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.205022] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 67d8c43a-d3af-407b-847a-1c2c8c53dbfb/67d8c43a-d3af-407b-847a-1c2c8c53dbfb.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 780.205022] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-edb8cf19-9afe-4fa0-8afd-5966ae01652c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.210309] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 780.210309] env[62368]: value = "task-1198363" [ 780.210309] env[62368]: _type = "Task" [ 780.210309] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.220753] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.314530] env[62368]: DEBUG nova.network.neutron [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.345306] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198362, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.416246] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2ad89cf6-8779-45a5-9246-d34d7af70a27 tempest-ServerMetadataTestJSON-1823330661 tempest-ServerMetadataTestJSON-1823330661-project-member] Lock "9ce5c5a1-5c90-4f76-a511-d8f855b6bab4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.255s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.474935] env[62368]: DEBUG nova.network.neutron [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Updating instance_info_cache with network_info: [{"id": "e7892749-2fef-4fa0-ba5b-348c58477c93", "address": "fa:16:3e:6d:e2:3f", "network": {"id": "80628028-4adf-4b96-9fcb-e9d06c4b5148", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1932221090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f00de1923334a2999f52d0a374cd522", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7892749-2f", "ovs_interfaceid": "e7892749-2fef-4fa0-ba5b-348c58477c93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.489863] env[62368]: DEBUG oslo_vmware.api [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198360, 'name': RemoveSnapshot_Task, 'duration_secs': 0.764067} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.490522] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 780.491669] env[62368]: INFO nova.compute.manager [None req-b9223e2c-8b89-4701-a67c-b0eab8162404 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Took 16.98 seconds to snapshot the instance on the hypervisor. [ 780.511255] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198361, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51819} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.512351] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 300c6c32-b365-40c2-bc7d-6ce028f10476/300c6c32-b365-40c2-bc7d-6ce028f10476.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 780.512351] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.512351] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1e9f6af-5b3f-4fbd-b23f-a91e166f7905 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.523479] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 780.523479] env[62368]: value = "task-1198364" [ 780.523479] env[62368]: _type = "Task" [ 780.523479] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.544783] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198364, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.722836] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490171} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.723130] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 67d8c43a-d3af-407b-847a-1c2c8c53dbfb/67d8c43a-d3af-407b-847a-1c2c8c53dbfb.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 780.723359] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.723632] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc65ceae-254e-41a0-9846-8049dac4e654 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.732676] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 780.732676] env[62368]: value = "task-1198365" [ 780.732676] env[62368]: _type = "Task" [ 780.732676] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.746585] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.823279] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Successfully updated port: 9897441b-2e19-4f96-8124-6aae0b31e308 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.843403] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198362, 'name': ReconfigVM_Task, 'duration_secs': 0.770123} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.844538] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 3f1af54b-392f-432a-9ffa-a133da428f94/3f1af54b-392f-432a-9ffa-a133da428f94.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 780.844855] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8abe905b-79c2-4a54-81f7-c75970aa03cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.854552] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 780.854552] env[62368]: value = "task-1198366" [ 780.854552] env[62368]: _type = "Task" [ 780.854552] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.864772] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198366, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.927848] env[62368]: DEBUG nova.network.neutron [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updating instance_info_cache with network_info: [{"id": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "address": "fa:16:3e:ba:65:a2", "network": {"id": "831eece9-e402-442b-b933-93e089c864be", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1971247571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd4cd004b0d54bb49a88722fd1312249", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3bbc3d-86", "ovs_interfaceid": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.948424] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c947d8-50ac-4060-a47e-a4d8453efbab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.957625] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdbd570-3a32-4abb-be1c-c54b7856dae8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.991134] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Releasing lock "refresh_cache-0f1bff84-1fb9-471a-b685-7d527b376b0a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.991382] env[62368]: DEBUG nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Instance network_info: |[{"id": "e7892749-2fef-4fa0-ba5b-348c58477c93", "address": "fa:16:3e:6d:e2:3f", "network": {"id": "80628028-4adf-4b96-9fcb-e9d06c4b5148", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1932221090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f00de1923334a2999f52d0a374cd522", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7892749-2f", "ovs_interfaceid": "e7892749-2fef-4fa0-ba5b-348c58477c93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 780.992169] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7432ef4b-be0b-4f7c-a746-2978c8bf2ef6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.997249] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:e2:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cd098b1c-636f-492d-b5ae-037cb0cae454', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7892749-2fef-4fa0-ba5b-348c58477c93', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.004486] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Creating folder: Project (8f00de1923334a2999f52d0a374cd522). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.007009] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c87d8fdb-a8a8-421e-9198-fcd3025c13e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.018042] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013ae34b-03bb-459e-a1d5-33956435f39e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.032855] env[62368]: DEBUG nova.compute.provider_tree [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.041533] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Created folder: Project (8f00de1923334a2999f52d0a374cd522) in parent group-v259706. [ 781.041533] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Creating folder: Instances. Parent ref: group-v259778. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.041732] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c806dfac-0d46-4f94-9239-f8af7b2efd1b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.050502] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198364, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.174644} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.050898] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.051811] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf892912-928b-4670-a9e3-f26c666fed24 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.056310] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Created folder: Instances in parent group-v259778. [ 781.056797] env[62368]: DEBUG oslo.service.loopingcall [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.057581] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 781.058229] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-316e06dc-75d8-4cc2-9b0b-3307586abad1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.098638] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 300c6c32-b365-40c2-bc7d-6ce028f10476/300c6c32-b365-40c2-bc7d-6ce028f10476.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.099524] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d754ccce-ac6e-4efd-a413-d3a1b9611614 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.116488] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.116488] env[62368]: value = "task-1198369" [ 781.116488] env[62368]: _type = "Task" [ 781.116488] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.123165] env[62368]: DEBUG nova.compute.manager [req-2db7f5b8-1513-4dc0-b814-9532a8bf440a req-e29e30f5-d915-4e75-9a30-3183b3f9554a service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Received event network-vif-plugged-9897441b-2e19-4f96-8124-6aae0b31e308 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.123440] env[62368]: DEBUG oslo_concurrency.lockutils [req-2db7f5b8-1513-4dc0-b814-9532a8bf440a req-e29e30f5-d915-4e75-9a30-3183b3f9554a service nova] Acquiring lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.123695] env[62368]: DEBUG oslo_concurrency.lockutils [req-2db7f5b8-1513-4dc0-b814-9532a8bf440a req-e29e30f5-d915-4e75-9a30-3183b3f9554a service nova] Lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.123896] env[62368]: DEBUG oslo_concurrency.lockutils [req-2db7f5b8-1513-4dc0-b814-9532a8bf440a req-e29e30f5-d915-4e75-9a30-3183b3f9554a service nova] Lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.124114] env[62368]: DEBUG nova.compute.manager [req-2db7f5b8-1513-4dc0-b814-9532a8bf440a req-e29e30f5-d915-4e75-9a30-3183b3f9554a service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] No waiting events found dispatching network-vif-plugged-9897441b-2e19-4f96-8124-6aae0b31e308 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 781.124294] env[62368]: WARNING nova.compute.manager [req-2db7f5b8-1513-4dc0-b814-9532a8bf440a req-e29e30f5-d915-4e75-9a30-3183b3f9554a service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Received unexpected event network-vif-plugged-9897441b-2e19-4f96-8124-6aae0b31e308 for instance with vm_state building and task_state spawning. [ 781.126846] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 781.126846] env[62368]: value = "task-1198370" [ 781.126846] env[62368]: _type = "Task" [ 781.126846] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.134423] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198369, 'name': CreateVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.140930] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198370, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.217604] env[62368]: DEBUG nova.compute.manager [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Received event network-vif-plugged-e7892749-2fef-4fa0-ba5b-348c58477c93 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.217841] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Acquiring lock "0f1bff84-1fb9-471a-b685-7d527b376b0a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.218577] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Lock "0f1bff84-1fb9-471a-b685-7d527b376b0a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.218800] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Lock "0f1bff84-1fb9-471a-b685-7d527b376b0a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.219036] env[62368]: DEBUG nova.compute.manager [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] No waiting events found dispatching network-vif-plugged-e7892749-2fef-4fa0-ba5b-348c58477c93 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 781.219200] env[62368]: WARNING nova.compute.manager [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Received unexpected event network-vif-plugged-e7892749-2fef-4fa0-ba5b-348c58477c93 for instance with vm_state building and task_state spawning. [ 781.219369] env[62368]: DEBUG nova.compute.manager [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Received event network-changed-e7892749-2fef-4fa0-ba5b-348c58477c93 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.219524] env[62368]: DEBUG nova.compute.manager [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Refreshing instance network info cache due to event network-changed-e7892749-2fef-4fa0-ba5b-348c58477c93. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 781.219716] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Acquiring lock "refresh_cache-0f1bff84-1fb9-471a-b685-7d527b376b0a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.219855] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Acquired lock "refresh_cache-0f1bff84-1fb9-471a-b685-7d527b376b0a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.220017] env[62368]: DEBUG nova.network.neutron [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Refreshing network info cache for port e7892749-2fef-4fa0-ba5b-348c58477c93 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 781.248533] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.291475} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.248799] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.249626] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04e7ab0-28cd-475f-931a-40207f5af8ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.274346] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 67d8c43a-d3af-407b-847a-1c2c8c53dbfb/67d8c43a-d3af-407b-847a-1c2c8c53dbfb.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.275672] env[62368]: DEBUG nova.network.neutron [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.276932] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae98e1a0-165c-49c5-84ff-fbf7f18ff855 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.300718] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 781.300718] env[62368]: value = "task-1198371" [ 781.300718] env[62368]: _type = "Task" [ 781.300718] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.313257] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198371, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.330547] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "refresh_cache-8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.330692] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "refresh_cache-8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.330861] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.368230] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198366, 'name': Rename_Task, 'duration_secs': 0.370361} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.368536] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.368815] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ca855f7-c4d4-4f2d-8347-6cd98d74845e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.376649] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 781.376649] env[62368]: value = "task-1198372" [ 781.376649] env[62368]: _type = "Task" [ 781.376649] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.387074] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198372, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.430565] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Releasing lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.433245] env[62368]: DEBUG nova.compute.manager [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 781.434421] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0922f41e-9f0f-4ed9-88c3-0411dcbdfd25 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.543133] env[62368]: DEBUG nova.scheduler.client.report [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.628089] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198369, 'name': CreateVM_Task, 'duration_secs': 0.427937} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.631548] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 781.634079] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.634079] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.634079] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 781.634079] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fa6ae14-958f-4755-bfbb-b32be4815405 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.638963] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198370, 'name': ReconfigVM_Task, 'duration_secs': 0.445851} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.644021] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 300c6c32-b365-40c2-bc7d-6ce028f10476/300c6c32-b365-40c2-bc7d-6ce028f10476.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.644021] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 781.644021] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5243c99d-1530-a492-dff7-219ee7e42507" [ 781.644021] env[62368]: _type = "Task" [ 781.644021] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.644021] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f552034f-bd36-407d-890c-cad187c166be {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.652275] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5243c99d-1530-a492-dff7-219ee7e42507, 'name': SearchDatastore_Task, 'duration_secs': 0.010317} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.654560] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.654560] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.654560] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.654560] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.654560] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.655046] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 781.655046] env[62368]: value = "task-1198373" [ 781.655046] env[62368]: _type = "Task" [ 781.655046] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.655046] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a467010-13b9-4dcf-9778-1b1c40330cd0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.668219] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198373, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.668832] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.669031] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 781.669852] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-254c7327-d79e-4025-9817-4b494f7937a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.675551] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 781.675551] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522fff78-3d30-21ac-63e3-6abdeb96cec9" [ 781.675551] env[62368]: _type = "Task" [ 781.675551] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.686859] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522fff78-3d30-21ac-63e3-6abdeb96cec9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.726132] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "d865d433-2341-4f6c-b840-609530ae2e51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.726428] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "d865d433-2341-4f6c-b840-609530ae2e51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.793450] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Releasing lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.793981] env[62368]: DEBUG nova.compute.manager [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Inject network info {{(pid=62368) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 781.794568] env[62368]: DEBUG nova.compute.manager [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] network_info to inject: |[{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 781.800344] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Reconfiguring VM instance to set the machine id {{(pid=62368) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 781.800808] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6028f9a6-3fc8-4a05-ae15-1d36ed63b454 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.833226] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198371, 'name': ReconfigVM_Task, 'duration_secs': 0.33485} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.833226] env[62368]: DEBUG oslo_vmware.api [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 781.833226] env[62368]: value = "task-1198374" [ 781.833226] env[62368]: _type = "Task" [ 781.833226] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.833226] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 67d8c43a-d3af-407b-847a-1c2c8c53dbfb/67d8c43a-d3af-407b-847a-1c2c8c53dbfb.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.833226] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3275f980-05db-49fa-8ce4-b4ac44e016e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.852309] env[62368]: DEBUG oslo_vmware.api [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.853958] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 781.853958] env[62368]: value = "task-1198375" [ 781.853958] env[62368]: _type = "Task" [ 781.853958] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.863460] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198375, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.887774] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198372, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.896505] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.049160] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.195s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.061572] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.264s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.063261] env[62368]: INFO nova.compute.claims [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.089592] env[62368]: INFO nova.scheduler.client.report [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Deleted allocations for instance b393f654-0611-493e-a89f-445dfd3f8c13 [ 782.126910] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Updating instance_info_cache with network_info: [{"id": "9897441b-2e19-4f96-8124-6aae0b31e308", "address": "fa:16:3e:0c:b9:2f", "network": {"id": "6f399836-27f3-47f7-af4b-e5dbeceee516", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1591352226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23e257a2707d4c5cb3062081be3c0546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9897441b-2e", "ovs_interfaceid": "9897441b-2e19-4f96-8124-6aae0b31e308", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.172620] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198373, 'name': Rename_Task, 'duration_secs': 0.170073} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.173699] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 782.173964] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1bbcf1e-7ad1-4c1c-86a2-b2b8f8c17743 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.182661] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 782.182661] env[62368]: value = "task-1198376" [ 782.182661] env[62368]: _type = "Task" [ 782.182661] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.186799] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522fff78-3d30-21ac-63e3-6abdeb96cec9, 'name': SearchDatastore_Task, 'duration_secs': 0.015639} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.191477] env[62368]: DEBUG nova.network.neutron [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Updated VIF entry in instance network info cache for port e7892749-2fef-4fa0-ba5b-348c58477c93. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 782.191837] env[62368]: DEBUG nova.network.neutron [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Updating instance_info_cache with network_info: [{"id": "e7892749-2fef-4fa0-ba5b-348c58477c93", "address": "fa:16:3e:6d:e2:3f", "network": {"id": "80628028-4adf-4b96-9fcb-e9d06c4b5148", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1932221090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f00de1923334a2999f52d0a374cd522", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7892749-2f", "ovs_interfaceid": "e7892749-2fef-4fa0-ba5b-348c58477c93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.192975] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c5494cb-d564-41f8-8b10-cbc919b310a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.201158] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198376, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.202498] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 782.202498] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52a758ce-d0e4-e0b9-2bed-dcb27dabe596" [ 782.202498] env[62368]: _type = "Task" [ 782.202498] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.211500] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a758ce-d0e4-e0b9-2bed-dcb27dabe596, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.294223] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.294554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.294849] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.295129] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.295320] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.297583] env[62368]: INFO nova.compute.manager [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Terminating instance [ 782.299501] env[62368]: DEBUG nova.compute.manager [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.299695] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 782.300592] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e569e24-c138-4ea1-b8dc-2306908154dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.311109] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 782.311373] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5863fb81-e5f6-4556-aabb-67c6487eb02c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.320122] env[62368]: DEBUG oslo_vmware.api [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 782.320122] env[62368]: value = "task-1198377" [ 782.320122] env[62368]: _type = "Task" [ 782.320122] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.332897] env[62368]: DEBUG oslo_vmware.api [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198377, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.346286] env[62368]: DEBUG oslo_vmware.api [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198374, 'name': ReconfigVM_Task, 'duration_secs': 0.15615} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.346286] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1de5fa70-8dfe-424c-b51d-6431e95e4764 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Reconfigured VM instance to set the machine id {{(pid=62368) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 782.365641] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198375, 'name': Rename_Task, 'duration_secs': 0.220617} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.365945] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 782.366418] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a04d10e0-1634-420a-93b4-f9208c878ca5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.375070] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 782.375070] env[62368]: value = "task-1198378" [ 782.375070] env[62368]: _type = "Task" [ 782.375070] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.392208] env[62368]: DEBUG oslo_vmware.api [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198372, 'name': PowerOnVM_Task, 'duration_secs': 0.801639} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.392446] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198378, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.392691] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 782.392915] env[62368]: INFO nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Took 10.17 seconds to spawn the instance on the hypervisor. [ 782.393114] env[62368]: DEBUG nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.393995] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891b3b81-043b-4342-a161-86053c7abbaa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.452577] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0315fe-1042-4cd4-b459-738571e49d62 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.463899] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Doing hard reboot of VM {{(pid=62368) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 782.464225] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0b0d6102-f9fa-445b-acae-db43ccbf060c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.477813] env[62368]: DEBUG oslo_vmware.api [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 782.477813] env[62368]: value = "task-1198379" [ 782.477813] env[62368]: _type = "Task" [ 782.477813] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.488038] env[62368]: DEBUG oslo_vmware.api [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198379, 'name': ResetVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.601368] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c2cd27d7-78bc-4c2d-b084-a673b06d97d8 tempest-InstanceActionsV221TestJSON-1387792481 tempest-InstanceActionsV221TestJSON-1387792481-project-member] Lock "b393f654-0611-493e-a89f-445dfd3f8c13" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.330s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.631404] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "refresh_cache-8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.631404] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Instance network_info: |[{"id": "9897441b-2e19-4f96-8124-6aae0b31e308", "address": "fa:16:3e:0c:b9:2f", "network": {"id": "6f399836-27f3-47f7-af4b-e5dbeceee516", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1591352226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23e257a2707d4c5cb3062081be3c0546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9897441b-2e", "ovs_interfaceid": "9897441b-2e19-4f96-8124-6aae0b31e308", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 782.631404] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:b9:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9897441b-2e19-4f96-8124-6aae0b31e308', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.643445] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Creating folder: Project (23e257a2707d4c5cb3062081be3c0546). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 782.644189] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da9c12ab-f736-4f0c-a41b-7bb3ecb4236f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.660601] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Created folder: Project (23e257a2707d4c5cb3062081be3c0546) in parent group-v259706. [ 782.660601] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Creating folder: Instances. Parent ref: group-v259781. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 782.660601] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ac8ddeb-9547-41a4-bc36-39346848d8ee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.671471] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Created folder: Instances in parent group-v259781. [ 782.671771] env[62368]: DEBUG oslo.service.loopingcall [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.673352] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 782.673923] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f73321e-8320-4f7a-a6dc-dbc26e97290b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.697877] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Releasing lock "refresh_cache-0f1bff84-1fb9-471a-b685-7d527b376b0a" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.698315] env[62368]: DEBUG nova.compute.manager [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Received event network-changed-d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.698617] env[62368]: DEBUG nova.compute.manager [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Refreshing instance network info cache due to event network-changed-d98e888b-22dd-483e-999f-4e91c0e15ab6. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.699064] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Acquiring lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.699257] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Acquired lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.699560] env[62368]: DEBUG nova.network.neutron [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Refreshing network info cache for port d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.707446] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.707446] env[62368]: value = "task-1198382" [ 782.707446] env[62368]: _type = "Task" [ 782.707446] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.717956] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198376, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.733548] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a758ce-d0e4-e0b9-2bed-dcb27dabe596, 'name': SearchDatastore_Task, 'duration_secs': 0.023339} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.738230] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.738230] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 0f1bff84-1fb9-471a-b685-7d527b376b0a/0f1bff84-1fb9-471a-b685-7d527b376b0a.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 782.738230] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198382, 'name': CreateVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.738834] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65fc9a8e-7f24-48b9-b87a-f9eeec852333 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.749698] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 782.749698] env[62368]: value = "task-1198383" [ 782.749698] env[62368]: _type = "Task" [ 782.749698] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.761586] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198383, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.834482] env[62368]: DEBUG oslo_vmware.api [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198377, 'name': PowerOffVM_Task, 'duration_secs': 0.216988} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.835485] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 782.835485] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 782.835485] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0309f4a2-f947-4d1c-8d2c-e6d356a810fb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.889041] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198378, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.916781] env[62368]: INFO nova.compute.manager [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Took 37.87 seconds to build instance. [ 782.923780] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 782.924183] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 782.924427] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Deleting the datastore file [datastore2] 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.924745] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94cc27e5-d0e9-44c1-99c6-14e0f0e6b466 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.938062] env[62368]: DEBUG oslo_vmware.api [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for the task: (returnval){ [ 782.938062] env[62368]: value = "task-1198385" [ 782.938062] env[62368]: _type = "Task" [ 782.938062] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.950153] env[62368]: DEBUG oslo_vmware.api [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.988621] env[62368]: DEBUG oslo_vmware.api [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198379, 'name': ResetVM_Task, 'duration_secs': 0.10136} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.988996] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Did hard reboot of VM {{(pid=62368) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 782.989222] env[62368]: DEBUG nova.compute.manager [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.990302] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4b9cb8-689f-4bef-981f-3b82653fceeb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.183702] env[62368]: DEBUG nova.compute.manager [req-63505b8f-2b5a-4af0-9110-395356bc9a42 req-449d6078-48fc-416b-ae35-791304928a48 service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Received event network-changed-9897441b-2e19-4f96-8124-6aae0b31e308 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.183950] env[62368]: DEBUG nova.compute.manager [req-63505b8f-2b5a-4af0-9110-395356bc9a42 req-449d6078-48fc-416b-ae35-791304928a48 service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Refreshing instance network info cache due to event network-changed-9897441b-2e19-4f96-8124-6aae0b31e308. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 783.185547] env[62368]: DEBUG oslo_concurrency.lockutils [req-63505b8f-2b5a-4af0-9110-395356bc9a42 req-449d6078-48fc-416b-ae35-791304928a48 service nova] Acquiring lock "refresh_cache-8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.185547] env[62368]: DEBUG oslo_concurrency.lockutils [req-63505b8f-2b5a-4af0-9110-395356bc9a42 req-449d6078-48fc-416b-ae35-791304928a48 service nova] Acquired lock "refresh_cache-8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.185547] env[62368]: DEBUG nova.network.neutron [req-63505b8f-2b5a-4af0-9110-395356bc9a42 req-449d6078-48fc-416b-ae35-791304928a48 service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Refreshing network info cache for port 9897441b-2e19-4f96-8124-6aae0b31e308 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 783.203258] env[62368]: DEBUG oslo_vmware.api [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198376, 'name': PowerOnVM_Task, 'duration_secs': 0.563224} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.208103] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 783.208366] env[62368]: INFO nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Took 13.47 seconds to spawn the instance on the hypervisor. [ 783.208600] env[62368]: DEBUG nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 783.212313] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2271e432-13e7-4833-9d5f-00a84909be26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.243937] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198382, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.264893] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198383, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.387995] env[62368]: DEBUG oslo_vmware.api [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198378, 'name': PowerOnVM_Task, 'duration_secs': 0.540043} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.393753] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 783.393753] env[62368]: INFO nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Took 8.62 seconds to spawn the instance on the hypervisor. [ 783.393753] env[62368]: DEBUG nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 783.395410] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6051cf6b-eec6-4836-b02e-c0e7a122da1f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.420208] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5ea3c659-ba93-4163-82af-0ae72a69c857 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "3f1af54b-392f-432a-9ffa-a133da428f94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.169s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.456265] env[62368]: DEBUG oslo_vmware.api [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Task: {'id': task-1198385, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323047} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.458869] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 783.459078] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 783.459472] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 783.459472] env[62368]: INFO nova.compute.manager [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Took 1.16 seconds to destroy the instance on the hypervisor. [ 783.459681] env[62368]: DEBUG oslo.service.loopingcall [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.460431] env[62368]: DEBUG nova.compute.manager [-] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.462117] env[62368]: DEBUG nova.network.neutron [-] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 783.507038] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bb12ef06-6cb9-41f9-b971-b87f074b25be tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.464s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.512571] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d27d185-cf11-4773-a5e2-e42e3e00caef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.527256] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64929ac1-4874-4d88-b6e2-55212b62e60d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.560688] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f8fe77-c11e-4b58-af75-c14c4c795443 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.569774] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e9ad22-6314-4642-9813-ef4bf9e0804a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.585121] env[62368]: DEBUG nova.compute.provider_tree [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.670936] env[62368]: DEBUG nova.network.neutron [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updated VIF entry in instance network info cache for port d98e888b-22dd-483e-999f-4e91c0e15ab6. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.671329] env[62368]: DEBUG nova.network.neutron [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [{"id": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "address": "fa:16:3e:96:10:1c", "network": {"id": "98932a5c-1cd9-475b-b435-58b4f36f4ada", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2087236777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31833b73a8724021bab7a6efedc7acfd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd98e888b-22", "ovs_interfaceid": "d98e888b-22dd-483e-999f-4e91c0e15ab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.729908] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198382, 'name': CreateVM_Task, 'duration_secs': 0.662905} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.730130] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 783.730814] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.730985] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.731344] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 783.731703] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-322cc7b7-d1a6-4806-8a56-753031c056ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.741339] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 783.741339] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5245fcb7-4503-a7bc-870f-3266e8d6cd16" [ 783.741339] env[62368]: _type = "Task" [ 783.741339] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.755091] env[62368]: INFO nova.compute.manager [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Took 42.30 seconds to build instance. [ 783.767536] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5245fcb7-4503-a7bc-870f-3266e8d6cd16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.771820] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198383, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651007} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.772120] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 0f1bff84-1fb9-471a-b685-7d527b376b0a/0f1bff84-1fb9-471a-b685-7d527b376b0a.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 783.772330] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.772883] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d448bbe1-9285-4b17-a4fe-4a6a6e0b46b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.782149] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 783.782149] env[62368]: value = "task-1198386" [ 783.782149] env[62368]: _type = "Task" [ 783.782149] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.795753] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.922453] env[62368]: INFO nova.compute.manager [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Took 33.87 seconds to build instance. [ 783.925925] env[62368]: DEBUG nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.089277] env[62368]: DEBUG nova.scheduler.client.report [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.174091] env[62368]: DEBUG oslo_concurrency.lockutils [req-1daf084b-ff35-47bd-ba28-58e4674de10e req-2c26d4b5-1df7-4c83-ad97-978960657acd service nova] Releasing lock "refresh_cache-8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.258020] env[62368]: DEBUG nova.network.neutron [req-63505b8f-2b5a-4af0-9110-395356bc9a42 req-449d6078-48fc-416b-ae35-791304928a48 service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Updated VIF entry in instance network info cache for port 9897441b-2e19-4f96-8124-6aae0b31e308. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 784.258020] env[62368]: DEBUG nova.network.neutron [req-63505b8f-2b5a-4af0-9110-395356bc9a42 req-449d6078-48fc-416b-ae35-791304928a48 service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Updating instance_info_cache with network_info: [{"id": "9897441b-2e19-4f96-8124-6aae0b31e308", "address": "fa:16:3e:0c:b9:2f", "network": {"id": "6f399836-27f3-47f7-af4b-e5dbeceee516", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1591352226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23e257a2707d4c5cb3062081be3c0546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9897441b-2e", "ovs_interfaceid": "9897441b-2e19-4f96-8124-6aae0b31e308", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.264370] env[62368]: DEBUG oslo_concurrency.lockutils [None req-86ca04c9-e46b-4dc6-a24f-e607945280ef tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.511s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.264756] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5245fcb7-4503-a7bc-870f-3266e8d6cd16, 'name': SearchDatastore_Task, 'duration_secs': 0.030977} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.266732] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.267075] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 784.267395] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.267602] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.267847] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 784.269009] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70bfe283-bd2a-4b86-938d-efa832cfb092 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.287970] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 784.288129] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 784.289171] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15bfb5b6-fa2c-4337-b3e9-9cf1051e9eeb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.294476] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.185889} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.295111] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.295955] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc73804f-8397-4995-8140-66d6facb65dd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.299855] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 784.299855] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52bee06c-9daa-e3b8-ec1c-ee7041880967" [ 784.299855] env[62368]: _type = "Task" [ 784.299855] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.323436] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 0f1bff84-1fb9-471a-b685-7d527b376b0a/0f1bff84-1fb9-471a-b685-7d527b376b0a.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.324733] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c41df340-00b5-40b1-b473-903dce3e4677 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.344403] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52bee06c-9daa-e3b8-ec1c-ee7041880967, 'name': SearchDatastore_Task, 'duration_secs': 0.017152} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.345956] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67b540ee-cbf5-4601-8ca1-d6db44775d43 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.350991] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 784.350991] env[62368]: value = "task-1198387" [ 784.350991] env[62368]: _type = "Task" [ 784.350991] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.352903] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 784.352903] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5212633b-1e01-9d3c-bc96-dda3a9263c3c" [ 784.352903] env[62368]: _type = "Task" [ 784.352903] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.367883] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5212633b-1e01-9d3c-bc96-dda3a9263c3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.372205] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198387, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.426661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1e7c737-dc08-45eb-a28a-a6a7662ccad8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.477s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.457526] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.599428] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.599932] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.603167] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.257s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.606304] env[62368]: INFO nova.compute.claims [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.766773] env[62368]: DEBUG oslo_concurrency.lockutils [req-63505b8f-2b5a-4af0-9110-395356bc9a42 req-449d6078-48fc-416b-ae35-791304928a48 service nova] Releasing lock "refresh_cache-8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.769499] env[62368]: DEBUG nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.885985] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198387, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.886324] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5212633b-1e01-9d3c-bc96-dda3a9263c3c, 'name': SearchDatastore_Task, 'duration_secs': 0.026534} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.886583] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.886993] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d/8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 784.887097] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d0ce45c2-997c-4a6b-9b96-01a14efa9259 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.899532] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 784.899532] env[62368]: value = "task-1198388" [ 784.899532] env[62368]: _type = "Task" [ 784.899532] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.919043] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.932511] env[62368]: DEBUG nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.116140] env[62368]: DEBUG nova.compute.utils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.118098] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.122053] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 785.189300] env[62368]: DEBUG nova.policy [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e17ec74ebf8d4c12accd841e4503b0df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23e257a2707d4c5cb3062081be3c0546', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.279714] env[62368]: DEBUG nova.network.neutron [-] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.283290] env[62368]: DEBUG nova.compute.manager [req-d9a8b16d-b12a-4fc1-8bf6-8a5c8e060a28 req-5a5d4e9e-f344-49d1-894e-fa2b4c245062 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Received event network-vif-deleted-d98e888b-22dd-483e-999f-4e91c0e15ab6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.283557] env[62368]: INFO nova.compute.manager [req-d9a8b16d-b12a-4fc1-8bf6-8a5c8e060a28 req-5a5d4e9e-f344-49d1-894e-fa2b4c245062 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Neutron deleted interface d98e888b-22dd-483e-999f-4e91c0e15ab6; detaching it from the instance and deleting it from the info cache [ 785.283856] env[62368]: DEBUG nova.network.neutron [req-d9a8b16d-b12a-4fc1-8bf6-8a5c8e060a28 req-5a5d4e9e-f344-49d1-894e-fa2b4c245062 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.312455] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.375761] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198387, 'name': ReconfigVM_Task, 'duration_secs': 0.60508} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.376348] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 0f1bff84-1fb9-471a-b685-7d527b376b0a/0f1bff84-1fb9-471a-b685-7d527b376b0a.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.377825] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bec50ca2-fe8d-4805-8441-5789a7dc4cbc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.391174] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 785.391174] env[62368]: value = "task-1198389" [ 785.391174] env[62368]: _type = "Task" [ 785.391174] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.407972] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198389, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.415380] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198388, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.463257] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.612689] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Successfully created port: 4b9518d9-5b60-41b5-a631-9eeba2186936 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.625823] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.667606] env[62368]: DEBUG nova.compute.manager [req-5fc65bd1-36db-4e19-b12d-d4ce6e850000 req-13ee00cc-0aed-4ed9-be7a-80685c2035f0 service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Received event network-changed-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.667726] env[62368]: DEBUG nova.compute.manager [req-5fc65bd1-36db-4e19-b12d-d4ce6e850000 req-13ee00cc-0aed-4ed9-be7a-80685c2035f0 service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Refreshing instance network info cache due to event network-changed-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 785.667986] env[62368]: DEBUG oslo_concurrency.lockutils [req-5fc65bd1-36db-4e19-b12d-d4ce6e850000 req-13ee00cc-0aed-4ed9-be7a-80685c2035f0 service nova] Acquiring lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.668481] env[62368]: DEBUG oslo_concurrency.lockutils [req-5fc65bd1-36db-4e19-b12d-d4ce6e850000 req-13ee00cc-0aed-4ed9-be7a-80685c2035f0 service nova] Acquired lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.668676] env[62368]: DEBUG nova.network.neutron [req-5fc65bd1-36db-4e19-b12d-d4ce6e850000 req-13ee00cc-0aed-4ed9-be7a-80685c2035f0 service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Refreshing network info cache for port 4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 785.786580] env[62368]: INFO nova.compute.manager [-] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Took 2.33 seconds to deallocate network for instance. [ 785.795455] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf0092f5-784c-42d6-b97e-5668cdd6b6fb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.814023] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6feb17e-892d-4a09-85f6-330b1dabf045 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.850921] env[62368]: DEBUG nova.compute.manager [req-d9a8b16d-b12a-4fc1-8bf6-8a5c8e060a28 req-5a5d4e9e-f344-49d1-894e-fa2b4c245062 service nova] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Detach interface failed, port_id=d98e888b-22dd-483e-999f-4e91c0e15ab6, reason: Instance 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 785.911398] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198389, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.921635] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66412} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.921972] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d/8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 785.922174] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.923411] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fcde0fc8-7c77-4146-af85-c613bf9f6425 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.931366] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 785.931366] env[62368]: value = "task-1198390" [ 785.931366] env[62368]: _type = "Task" [ 785.931366] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.943551] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198390, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.994329] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778a9cbb-a18e-437b-9e20-a6f5c9630b9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.002077] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d320ebf-8b52-4e05-9bcc-c8d50bb696ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.040719] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbcc3af-861b-4ce2-8409-3989ffa5290d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.043686] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "300c6c32-b365-40c2-bc7d-6ce028f10476" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.043918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.044133] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.044316] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.044484] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.046785] env[62368]: INFO nova.compute.manager [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Terminating instance [ 786.051674] env[62368]: DEBUG nova.compute.manager [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 786.051910] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 786.053190] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de4317d-4445-4fce-a13c-385e59aad4a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.059561] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa17f1e-4458-48a3-93cf-494f25aaff99 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.074635] env[62368]: DEBUG nova.compute.provider_tree [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.080270] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 786.080270] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43416f54-5371-4251-b6a9-72c5299634ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.089788] env[62368]: DEBUG oslo_vmware.api [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 786.089788] env[62368]: value = "task-1198391" [ 786.089788] env[62368]: _type = "Task" [ 786.089788] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.094671] env[62368]: INFO nova.compute.manager [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Rebuilding instance [ 786.104230] env[62368]: DEBUG oslo_vmware.api [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198391, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.144452] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.144560] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.145618] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.145618] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.145618] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.148094] env[62368]: INFO nova.compute.manager [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Terminating instance [ 786.149919] env[62368]: DEBUG nova.compute.manager [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 786.150139] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 786.150976] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0da70a9-6de8-4a92-b4e1-e78eeb4401b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.165351] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 786.165351] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bfabf99f-5244-4a66-a836-d03f0da6a850 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.173088] env[62368]: DEBUG oslo_vmware.api [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 786.173088] env[62368]: value = "task-1198392" [ 786.173088] env[62368]: _type = "Task" [ 786.173088] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.178559] env[62368]: DEBUG nova.compute.manager [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 786.182508] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed8f65d-5041-4680-86c6-52b1ec99a92d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.192514] env[62368]: DEBUG oslo_vmware.api [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198392, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.296416] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.407765] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198389, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.442137] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198390, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076096} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.442421] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.447018] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b179c1-9cd2-4e0b-bbb7-e745a0588936 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.469065] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d/8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.471778] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea807099-a566-4167-a476-a351aaf2943a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.493531] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 786.493531] env[62368]: value = "task-1198393" [ 786.493531] env[62368]: _type = "Task" [ 786.493531] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.504786] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198393, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.580278] env[62368]: DEBUG nova.scheduler.client.report [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.603593] env[62368]: DEBUG oslo_vmware.api [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198391, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.648770] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 786.685109] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.685427] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.685594] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.686067] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.686274] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.686434] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.686649] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.686815] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.686985] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.687343] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.687575] env[62368]: DEBUG nova.virt.hardware [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.688832] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda85fc9-c91a-48df-bc89-0475d8944c07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.702885] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 786.703215] env[62368]: DEBUG oslo_vmware.api [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198392, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.707385] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f088030-bada-4079-9b1b-0793102b39b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.714414] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75bfa50-7f65-4f88-8af4-85adad9cd226 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.737993] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 786.737993] env[62368]: value = "task-1198394" [ 786.737993] env[62368]: _type = "Task" [ 786.737993] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.748121] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198394, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.834178] env[62368]: DEBUG nova.network.neutron [req-5fc65bd1-36db-4e19-b12d-d4ce6e850000 req-13ee00cc-0aed-4ed9-be7a-80685c2035f0 service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updated VIF entry in instance network info cache for port 4f3bbc3d-86d4-442f-9f71-dc40a130ce4c. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 786.834629] env[62368]: DEBUG nova.network.neutron [req-5fc65bd1-36db-4e19-b12d-d4ce6e850000 req-13ee00cc-0aed-4ed9-be7a-80685c2035f0 service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updating instance_info_cache with network_info: [{"id": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "address": "fa:16:3e:ba:65:a2", "network": {"id": "831eece9-e402-442b-b933-93e089c864be", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1971247571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd4cd004b0d54bb49a88722fd1312249", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e028024-a9c1-4cae-8849-ea770a7ae0e4", "external-id": "nsx-vlan-transportzone-919", "segmentation_id": 919, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f3bbc3d-86", "ovs_interfaceid": "4f3bbc3d-86d4-442f-9f71-dc40a130ce4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.909154] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198389, 'name': Rename_Task, 'duration_secs': 1.330461} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.910117] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 786.910351] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-444d419a-98b3-454c-88fd-12941f850041 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.920164] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 786.920164] env[62368]: value = "task-1198395" [ 786.920164] env[62368]: _type = "Task" [ 786.920164] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.930655] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.006845] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.088429] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.089686] env[62368]: DEBUG nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.092968] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.715s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.096662] env[62368]: DEBUG nova.objects.instance [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lazy-loading 'resources' on Instance uuid 085ebe93-aa24-4626-94fe-241c4297e4db {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 787.111928] env[62368]: DEBUG oslo_vmware.api [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198391, 'name': PowerOffVM_Task, 'duration_secs': 0.787981} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.112123] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 787.112413] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 787.112575] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18c3d696-e736-447b-b28b-fcf739f47f34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.189184] env[62368]: DEBUG oslo_vmware.api [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198392, 'name': PowerOffVM_Task, 'duration_secs': 0.612421} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.189184] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 787.189326] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 787.193024] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a31aa32-84d6-4025-a620-7bc62a60e849 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.249259] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198394, 'name': PowerOffVM_Task, 'duration_secs': 0.22651} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.249259] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 787.249259] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 787.250341] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127176c5-da11-42a8-aafc-e5f330002c8a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.259153] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 787.259433] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-902a6e7d-5257-4dc7-9255-ff4ba0613845 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.276066] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 787.276400] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 787.276566] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Deleting the datastore file [datastore1] 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.277895] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21399359-ee9a-42b5-9b8d-a3511aa27caf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.279919] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 787.280150] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 787.280330] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Deleting the datastore file [datastore1] 300c6c32-b365-40c2-bc7d-6ce028f10476 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.280590] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0ec50f5-46ac-44c6-be42-11acaec3dab9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.288268] env[62368]: DEBUG oslo_vmware.api [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for the task: (returnval){ [ 787.288268] env[62368]: value = "task-1198399" [ 787.288268] env[62368]: _type = "Task" [ 787.288268] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.293780] env[62368]: DEBUG oslo_vmware.api [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for the task: (returnval){ [ 787.293780] env[62368]: value = "task-1198400" [ 787.293780] env[62368]: _type = "Task" [ 787.293780] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.300985] env[62368]: DEBUG oslo_vmware.api [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198399, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.309834] env[62368]: DEBUG oslo_vmware.api [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.337986] env[62368]: DEBUG oslo_concurrency.lockutils [req-5fc65bd1-36db-4e19-b12d-d4ce6e850000 req-13ee00cc-0aed-4ed9-be7a-80685c2035f0 service nova] Releasing lock "refresh_cache-0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.349252] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 787.349252] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 787.349252] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleting the datastore file [datastore1] 3f1af54b-392f-432a-9ffa-a133da428f94 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.349252] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa6856e3-7147-42d9-95ee-e9db142f6070 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.359705] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 787.359705] env[62368]: value = "task-1198401" [ 787.359705] env[62368]: _type = "Task" [ 787.359705] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.372227] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.439918] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198395, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.482489] env[62368]: DEBUG nova.compute.manager [req-547c0180-cc7c-44f6-9107-cd2f551bc655 req-cfd99190-1a10-4f07-8867-8c33d97987ce service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Received event network-vif-plugged-4b9518d9-5b60-41b5-a631-9eeba2186936 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.482841] env[62368]: DEBUG oslo_concurrency.lockutils [req-547c0180-cc7c-44f6-9107-cd2f551bc655 req-cfd99190-1a10-4f07-8867-8c33d97987ce service nova] Acquiring lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.483189] env[62368]: DEBUG oslo_concurrency.lockutils [req-547c0180-cc7c-44f6-9107-cd2f551bc655 req-cfd99190-1a10-4f07-8867-8c33d97987ce service nova] Lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.483457] env[62368]: DEBUG oslo_concurrency.lockutils [req-547c0180-cc7c-44f6-9107-cd2f551bc655 req-cfd99190-1a10-4f07-8867-8c33d97987ce service nova] Lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.483739] env[62368]: DEBUG nova.compute.manager [req-547c0180-cc7c-44f6-9107-cd2f551bc655 req-cfd99190-1a10-4f07-8867-8c33d97987ce service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] No waiting events found dispatching network-vif-plugged-4b9518d9-5b60-41b5-a631-9eeba2186936 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 787.483998] env[62368]: WARNING nova.compute.manager [req-547c0180-cc7c-44f6-9107-cd2f551bc655 req-cfd99190-1a10-4f07-8867-8c33d97987ce service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Received unexpected event network-vif-plugged-4b9518d9-5b60-41b5-a631-9eeba2186936 for instance with vm_state building and task_state spawning. [ 787.509519] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198393, 'name': ReconfigVM_Task, 'duration_secs': 0.643472} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.509919] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d/8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.510617] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87347937-0e9e-463a-a471-359393860822 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.520770] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 787.520770] env[62368]: value = "task-1198402" [ 787.520770] env[62368]: _type = "Task" [ 787.520770] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.533310] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198402, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.598790] env[62368]: DEBUG nova.compute.utils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.601025] env[62368]: DEBUG nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.601025] env[62368]: DEBUG nova.network.neutron [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 787.672658] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Successfully updated port: 4b9518d9-5b60-41b5-a631-9eeba2186936 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.714545] env[62368]: DEBUG nova.policy [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '86d43d611b904b1c88e925db1e7f26ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '514286cc83654bf2a7a01d6c5df2b195', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.808685] env[62368]: DEBUG oslo_vmware.api [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Task: {'id': task-1198400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212088} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.811872] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.812129] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 787.812322] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 787.812502] env[62368]: INFO nova.compute.manager [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Took 1.76 seconds to destroy the instance on the hypervisor. [ 787.812759] env[62368]: DEBUG oslo.service.loopingcall [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.812983] env[62368]: DEBUG oslo_vmware.api [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Task: {'id': task-1198399, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195425} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.816187] env[62368]: DEBUG nova.compute.manager [-] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.816187] env[62368]: DEBUG nova.network.neutron [-] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.818335] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.818335] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 787.818335] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 787.818335] env[62368]: INFO nova.compute.manager [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Took 1.67 seconds to destroy the instance on the hypervisor. [ 787.818565] env[62368]: DEBUG oslo.service.loopingcall [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.819123] env[62368]: DEBUG nova.compute.manager [-] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.819224] env[62368]: DEBUG nova.network.neutron [-] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.873150] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184295} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.873432] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.873622] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 787.873867] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 787.934387] env[62368]: DEBUG oslo_vmware.api [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198395, 'name': PowerOnVM_Task, 'duration_secs': 0.60873} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.934387] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 787.934387] env[62368]: INFO nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Took 10.63 seconds to spawn the instance on the hypervisor. [ 787.934387] env[62368]: DEBUG nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.935360] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089ab35c-42dd-4c56-a226-d4bb9779ec32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.005017] env[62368]: DEBUG nova.compute.manager [req-70ccb561-1805-4c0c-b175-e762eb980c45 req-17c028de-7f10-4981-8e14-a4a203ae2105 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-changed-6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.005088] env[62368]: DEBUG nova.compute.manager [req-70ccb561-1805-4c0c-b175-e762eb980c45 req-17c028de-7f10-4981-8e14-a4a203ae2105 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Refreshing instance network info cache due to event network-changed-6a1f27ab-91ba-4185-9fdb-c5e31d622ed6. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.005434] env[62368]: DEBUG oslo_concurrency.lockutils [req-70ccb561-1805-4c0c-b175-e762eb980c45 req-17c028de-7f10-4981-8e14-a4a203ae2105 service nova] Acquiring lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.005434] env[62368]: DEBUG oslo_concurrency.lockutils [req-70ccb561-1805-4c0c-b175-e762eb980c45 req-17c028de-7f10-4981-8e14-a4a203ae2105 service nova] Acquired lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.007015] env[62368]: DEBUG nova.network.neutron [req-70ccb561-1805-4c0c-b175-e762eb980c45 req-17c028de-7f10-4981-8e14-a4a203ae2105 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Refreshing network info cache for port 6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 788.034876] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198402, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.065764] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41cdc3b-9316-4d39-9d87-e1fe42512557 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.077386] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afe5298-3755-4431-9908-16d247f19d3b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.112722] env[62368]: DEBUG nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.116288] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59e5dc1-75ad-4a55-98cc-96eff2e448d9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.125746] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a9c20d-8b5f-40e4-a961-848cc5c4879c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.141558] env[62368]: DEBUG nova.compute.provider_tree [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.177900] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "refresh_cache-87dc392f-7bba-44bb-ba38-2fd4e1c6129e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.178623] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "refresh_cache-87dc392f-7bba-44bb-ba38-2fd4e1c6129e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.178930] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.350403] env[62368]: DEBUG nova.network.neutron [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Successfully created port: 444a47e3-2f56-4895-8a8b-7e5605dbfaf1 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.459474] env[62368]: INFO nova.compute.manager [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Took 35.35 seconds to build instance. [ 788.535179] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198402, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.646513] env[62368]: DEBUG nova.scheduler.client.report [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.745710] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.814577] env[62368]: DEBUG nova.network.neutron [-] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.935132] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.935132] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.935132] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.935132] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.935132] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.935132] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.936049] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.936411] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.936742] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.937319] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.938061] env[62368]: DEBUG nova.virt.hardware [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.938956] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3debc4-330a-4a4a-95ce-00fb1cf25b1c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.945953] env[62368]: DEBUG nova.network.neutron [req-70ccb561-1805-4c0c-b175-e762eb980c45 req-17c028de-7f10-4981-8e14-a4a203ae2105 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updated VIF entry in instance network info cache for port 6a1f27ab-91ba-4185-9fdb-c5e31d622ed6. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.946276] env[62368]: DEBUG nova.network.neutron [req-70ccb561-1805-4c0c-b175-e762eb980c45 req-17c028de-7f10-4981-8e14-a4a203ae2105 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.958390] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0556c6e0-bf78-40d7-8f6a-550d5a7b4d01 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.964279] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dbe4c922-fe58-4bf7-829b-5f2e62fece46 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "0f1bff84-1fb9-471a-b685-7d527b376b0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.848s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.979135] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:ad:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c96365aa-f01b-493f-b516-a4e22b5b43e1', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.987674] env[62368]: DEBUG oslo.service.loopingcall [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.988311] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 788.988536] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de7ba2ac-44b8-48a0-8eba-c7b3aa5ebe63 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.009630] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.009630] env[62368]: value = "task-1198403" [ 789.009630] env[62368]: _type = "Task" [ 789.009630] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.010694] env[62368]: DEBUG nova.network.neutron [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Updating instance_info_cache with network_info: [{"id": "4b9518d9-5b60-41b5-a631-9eeba2186936", "address": "fa:16:3e:50:93:80", "network": {"id": "6f399836-27f3-47f7-af4b-e5dbeceee516", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1591352226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23e257a2707d4c5cb3062081be3c0546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b9518d9-5b", "ovs_interfaceid": "4b9518d9-5b60-41b5-a631-9eeba2186936", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.021116] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198403, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.032633] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198402, 'name': Rename_Task, 'duration_secs': 1.311905} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.033592] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 789.033863] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e5cfc2e-220a-455c-95e8-0ccfe2b40e6b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.043012] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 789.043012] env[62368]: value = "task-1198404" [ 789.043012] env[62368]: _type = "Task" [ 789.043012] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.053627] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198404, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.110320] env[62368]: DEBUG nova.network.neutron [-] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.128767] env[62368]: DEBUG nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.154657] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.062s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.157111] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.121s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.158662] env[62368]: INFO nova.compute.claims [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.172067] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:16:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f181c86a-16f3-4905-b3d9-783b0740eec7',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1115957553',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.172334] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.172629] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.172707] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.172848] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.173605] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.175546] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.175546] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.175546] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.175700] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.175915] env[62368]: DEBUG nova.virt.hardware [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.177149] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95a4ad8-44ee-437a-a26e-22d26f1335a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.190207] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe10d63-dd6b-49c4-b668-a03075eaddf3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.195402] env[62368]: INFO nova.scheduler.client.report [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Deleted allocations for instance 085ebe93-aa24-4626-94fe-241c4297e4db [ 789.317849] env[62368]: INFO nova.compute.manager [-] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Took 1.50 seconds to deallocate network for instance. [ 789.387128] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.387128] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.423871] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.424373] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.450614] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.450855] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.454758] env[62368]: DEBUG oslo_concurrency.lockutils [req-70ccb561-1805-4c0c-b175-e762eb980c45 req-17c028de-7f10-4981-8e14-a4a203ae2105 service nova] Releasing lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.516419] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "refresh_cache-87dc392f-7bba-44bb-ba38-2fd4e1c6129e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.516796] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Instance network_info: |[{"id": "4b9518d9-5b60-41b5-a631-9eeba2186936", "address": "fa:16:3e:50:93:80", "network": {"id": "6f399836-27f3-47f7-af4b-e5dbeceee516", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1591352226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23e257a2707d4c5cb3062081be3c0546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b9518d9-5b", "ovs_interfaceid": "4b9518d9-5b60-41b5-a631-9eeba2186936", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 789.517990] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:93:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b9518d9-5b60-41b5-a631-9eeba2186936', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.525211] env[62368]: DEBUG oslo.service.loopingcall [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.525976] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 789.526104] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37997200-d5a9-424e-863b-daa3ccb81dd8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.544719] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198403, 'name': CreateVM_Task, 'duration_secs': 0.37046} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.549605] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 789.549887] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.550045] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.550399] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.552110] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-055bfe97-398f-43c7-ad1c-05a48e4395cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.553601] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.553601] env[62368]: value = "task-1198405" [ 789.553601] env[62368]: _type = "Task" [ 789.553601] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.556826] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198404, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.561226] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 789.561226] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]528c5403-ef21-8062-8f79-fcb30c60eb63" [ 789.561226] env[62368]: _type = "Task" [ 789.561226] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.568078] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198405, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.573255] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528c5403-ef21-8062-8f79-fcb30c60eb63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.613571] env[62368]: INFO nova.compute.manager [-] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Took 1.80 seconds to deallocate network for instance. [ 789.683634] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a14bb49-be45-4af4-896b-3441e67d1f85 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "interface-0f1bff84-1fb9-471a-b685-7d527b376b0a-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.683867] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a14bb49-be45-4af4-896b-3441e67d1f85 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "interface-0f1bff84-1fb9-471a-b685-7d527b376b0a-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.684773] env[62368]: DEBUG nova.objects.instance [None req-3a14bb49-be45-4af4-896b-3441e67d1f85 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lazy-loading 'flavor' on Instance uuid 0f1bff84-1fb9-471a-b685-7d527b376b0a {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 789.705065] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8aa5a8ec-2d0a-4778-aa3b-075ed96f2462 tempest-FloatingIPsAssociationNegativeTestJSON-579986205 tempest-FloatingIPsAssociationNegativeTestJSON-579986205-project-member] Lock "085ebe93-aa24-4626-94fe-241c4297e4db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.317s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.824747] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.869062] env[62368]: DEBUG nova.compute.manager [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Received event network-changed-4b9518d9-5b60-41b5-a631-9eeba2186936 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.869387] env[62368]: DEBUG nova.compute.manager [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Refreshing instance network info cache due to event network-changed-4b9518d9-5b60-41b5-a631-9eeba2186936. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.869724] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] Acquiring lock "refresh_cache-87dc392f-7bba-44bb-ba38-2fd4e1c6129e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.870019] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] Acquired lock "refresh_cache-87dc392f-7bba-44bb-ba38-2fd4e1c6129e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.870402] env[62368]: DEBUG nova.network.neutron [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Refreshing network info cache for port 4b9518d9-5b60-41b5-a631-9eeba2186936 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.889152] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.060020] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198404, 'name': PowerOnVM_Task, 'duration_secs': 0.717616} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.060020] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 790.060020] env[62368]: INFO nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Took 10.24 seconds to spawn the instance on the hypervisor. [ 790.060020] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.062707] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eac9623-09ff-451f-8a16-64665dd56733 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.075029] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198405, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.091198] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528c5403-ef21-8062-8f79-fcb30c60eb63, 'name': SearchDatastore_Task, 'duration_secs': 0.015004} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.091198] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.091198] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.091198] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.091449] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.091569] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.092074] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76a31ed4-5a36-4333-9557-c5ed8fdb0037 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.105993] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.106275] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 790.107161] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34592849-7f94-4d58-9490-01c5e36f8185 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.116744] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 790.116744] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526f08a5-0bc3-e34f-70dc-36cce223dd20" [ 790.116744] env[62368]: _type = "Task" [ 790.116744] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.125069] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.134704] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526f08a5-0bc3-e34f-70dc-36cce223dd20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.191199] env[62368]: DEBUG nova.objects.instance [None req-3a14bb49-be45-4af4-896b-3441e67d1f85 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lazy-loading 'pci_requests' on Instance uuid 0f1bff84-1fb9-471a-b685-7d527b376b0a {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 790.420855] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.568398] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198405, 'name': CreateVM_Task, 'duration_secs': 0.667034} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.569817] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 790.570719] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41631a7b-e815-431c-b457-cea0163cb2a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.573935] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.574256] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.574788] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.576127] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73ee1ead-c89f-4639-b4f5-17182547b1eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.585087] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 790.585087] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52349096-dc45-7ac2-c67d-026fe9fc8739" [ 790.585087] env[62368]: _type = "Task" [ 790.585087] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.589757] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80a12de-9ce3-4388-b3f6-c547294a8c06 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.607334] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52349096-dc45-7ac2-c67d-026fe9fc8739, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.608249] env[62368]: INFO nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Took 37.45 seconds to build instance. [ 790.643672] env[62368]: DEBUG nova.network.neutron [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Successfully updated port: 444a47e3-2f56-4895-8a8b-7e5605dbfaf1 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.648674] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d399f17-ee8d-4cea-93e5-b32a52a9ea97 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.660194] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526f08a5-0bc3-e34f-70dc-36cce223dd20, 'name': SearchDatastore_Task, 'duration_secs': 0.013561} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.662487] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a8b90f3-5600-41e8-9387-e30cc277bff7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.665990] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962c2c6c-5f3e-4fc1-bd1c-39a88b58b71d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.675426] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 790.675426] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]525f3017-5856-2406-3ec1-184a6e4e94fb" [ 790.675426] env[62368]: _type = "Task" [ 790.675426] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.684462] env[62368]: DEBUG nova.compute.provider_tree [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.693488] env[62368]: DEBUG nova.objects.base [None req-3a14bb49-be45-4af4-896b-3441e67d1f85 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Object Instance<0f1bff84-1fb9-471a-b685-7d527b376b0a> lazy-loaded attributes: flavor,pci_requests {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 790.693703] env[62368]: DEBUG nova.network.neutron [None req-3a14bb49-be45-4af4-896b-3441e67d1f85 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 790.699711] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525f3017-5856-2406-3ec1-184a6e4e94fb, 'name': SearchDatastore_Task, 'duration_secs': 0.021197} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.700647] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.700647] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3f1af54b-392f-432a-9ffa-a133da428f94/3f1af54b-392f-432a-9ffa-a133da428f94.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 790.700800] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e431df39-9ca9-46ee-9af3-0af053d56d1d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.709327] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 790.709327] env[62368]: value = "task-1198406" [ 790.709327] env[62368]: _type = "Task" [ 790.709327] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.719093] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.864142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a14bb49-be45-4af4-896b-3441e67d1f85 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "interface-0f1bff84-1fb9-471a-b685-7d527b376b0a-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.179s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.929310] env[62368]: DEBUG nova.network.neutron [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Updated VIF entry in instance network info cache for port 4b9518d9-5b60-41b5-a631-9eeba2186936. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.929775] env[62368]: DEBUG nova.network.neutron [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Updating instance_info_cache with network_info: [{"id": "4b9518d9-5b60-41b5-a631-9eeba2186936", "address": "fa:16:3e:50:93:80", "network": {"id": "6f399836-27f3-47f7-af4b-e5dbeceee516", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1591352226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23e257a2707d4c5cb3062081be3c0546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b9518d9-5b", "ovs_interfaceid": "4b9518d9-5b60-41b5-a631-9eeba2186936", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.934510] env[62368]: DEBUG nova.compute.manager [req-d83fcd4f-4ecd-4318-a351-fdfa1580cf0a req-94fa487e-8ffd-4b7b-bccb-671a18bd3137 service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Received event network-vif-plugged-444a47e3-2f56-4895-8a8b-7e5605dbfaf1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.934882] env[62368]: DEBUG oslo_concurrency.lockutils [req-d83fcd4f-4ecd-4318-a351-fdfa1580cf0a req-94fa487e-8ffd-4b7b-bccb-671a18bd3137 service nova] Acquiring lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.935205] env[62368]: DEBUG oslo_concurrency.lockutils [req-d83fcd4f-4ecd-4318-a351-fdfa1580cf0a req-94fa487e-8ffd-4b7b-bccb-671a18bd3137 service nova] Lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.935665] env[62368]: DEBUG oslo_concurrency.lockutils [req-d83fcd4f-4ecd-4318-a351-fdfa1580cf0a req-94fa487e-8ffd-4b7b-bccb-671a18bd3137 service nova] Lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.935721] env[62368]: DEBUG nova.compute.manager [req-d83fcd4f-4ecd-4318-a351-fdfa1580cf0a req-94fa487e-8ffd-4b7b-bccb-671a18bd3137 service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] No waiting events found dispatching network-vif-plugged-444a47e3-2f56-4895-8a8b-7e5605dbfaf1 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.937046] env[62368]: WARNING nova.compute.manager [req-d83fcd4f-4ecd-4318-a351-fdfa1580cf0a req-94fa487e-8ffd-4b7b-bccb-671a18bd3137 service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Received unexpected event network-vif-plugged-444a47e3-2f56-4895-8a8b-7e5605dbfaf1 for instance with vm_state building and task_state spawning. [ 791.098201] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52349096-dc45-7ac2-c67d-026fe9fc8739, 'name': SearchDatastore_Task, 'duration_secs': 0.031078} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.101021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.101021] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.101021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.101021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.101021] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.101021] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fc0fc95-c283-44e9-abd3-41ca268ac269 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.117420] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.119513] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 791.120460] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-072fb1a0-d676-4100-b9ff-8adc91a543bc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.127861] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 791.127861] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]528f2502-2b0a-6851-f3e3-6067f86e9148" [ 791.127861] env[62368]: _type = "Task" [ 791.127861] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.137432] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528f2502-2b0a-6851-f3e3-6067f86e9148, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.142785] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.039s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.156330] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.156330] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.156330] env[62368]: DEBUG nova.network.neutron [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.193509] env[62368]: DEBUG nova.scheduler.client.report [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.224257] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198406, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.438328] env[62368]: DEBUG oslo_concurrency.lockutils [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] Releasing lock "refresh_cache-87dc392f-7bba-44bb-ba38-2fd4e1c6129e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.438328] env[62368]: DEBUG nova.compute.manager [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Received event network-vif-deleted-4f3bbc3d-86d4-442f-9f71-dc40a130ce4c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.438328] env[62368]: DEBUG nova.compute.manager [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Received event network-vif-deleted-59b16e71-e7dd-438f-812a-c369af2f64be {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.438328] env[62368]: DEBUG nova.compute.manager [req-9f512137-ebd3-4144-b3f6-cf03642535f0 req-b36420f2-410d-4a88-a36d-d6a692c8427d service nova] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Received event network-vif-deleted-64b32060-dc43-4d7e-88b2-84f2cb0e9055 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.642018] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528f2502-2b0a-6851-f3e3-6067f86e9148, 'name': SearchDatastore_Task, 'duration_secs': 0.064748} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.642939] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40d8c24c-2eb2-4fa4-9d0b-1988e3c5fe93 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.645690] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 791.657170] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 791.657170] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520c4517-8419-07a0-1eb7-9edd727ddc02" [ 791.657170] env[62368]: _type = "Task" [ 791.657170] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.668077] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520c4517-8419-07a0-1eb7-9edd727ddc02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.702027] env[62368]: DEBUG nova.network.neutron [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.703893] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.705181] env[62368]: DEBUG nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.709749] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.920s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.711034] env[62368]: INFO nova.compute.claims [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.727380] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198406, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.697231} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.729970] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3f1af54b-392f-432a-9ffa-a133da428f94/3f1af54b-392f-432a-9ffa-a133da428f94.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 791.730252] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.730529] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-16625799-c7b2-45d1-a9f2-9e68a0e38eb5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.742368] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 791.742368] env[62368]: value = "task-1198407" [ 791.742368] env[62368]: _type = "Task" [ 791.742368] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.762492] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.019525] env[62368]: DEBUG nova.network.neutron [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance_info_cache with network_info: [{"id": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "address": "fa:16:3e:98:a0:d2", "network": {"id": "34f156be-cab3-45a2-8b52-74dd4cd1e42d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2a054e975862463bb842479ad19393ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap444a47e3-2f", "ovs_interfaceid": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.177272] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520c4517-8419-07a0-1eb7-9edd727ddc02, 'name': SearchDatastore_Task, 'duration_secs': 0.040778} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.177689] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.178106] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 87dc392f-7bba-44bb-ba38-2fd4e1c6129e/87dc392f-7bba-44bb-ba38-2fd4e1c6129e.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 792.178458] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-168377d8-3d52-4d64-84ce-ded7af718d56 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.186326] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.187950] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 792.187950] env[62368]: value = "task-1198408" [ 792.187950] env[62368]: _type = "Task" [ 792.187950] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.197633] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.210471] env[62368]: DEBUG nova.compute.utils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 792.217224] env[62368]: DEBUG nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 792.217224] env[62368]: DEBUG nova.network.neutron [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 792.257962] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069645} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.258313] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.259145] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d1483c-c06a-4831-8967-2b02db647a5f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.287710] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 3f1af54b-392f-432a-9ffa-a133da428f94/3f1af54b-392f-432a-9ffa-a133da428f94.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.288076] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2cee58e-8098-4c89-a52c-70e493b4214e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.311572] env[62368]: DEBUG nova.policy [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c66c425264424e3eaa37c8058e62cb30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '026a3176abc34001ab3506445226777f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 792.314922] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 792.314922] env[62368]: value = "task-1198409" [ 792.314922] env[62368]: _type = "Task" [ 792.314922] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.328396] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198409, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.521238] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.521624] env[62368]: DEBUG nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Instance network_info: |[{"id": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "address": "fa:16:3e:98:a0:d2", "network": {"id": "34f156be-cab3-45a2-8b52-74dd4cd1e42d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2a054e975862463bb842479ad19393ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap444a47e3-2f", "ovs_interfaceid": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.522247] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:a0:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '444a47e3-2f56-4895-8a8b-7e5605dbfaf1', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.532116] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Creating folder: Project (514286cc83654bf2a7a01d6c5df2b195). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.532510] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6065bc9c-5b9a-4c6a-83df-13b6ade28a6a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.548070] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Created folder: Project (514286cc83654bf2a7a01d6c5df2b195) in parent group-v259706. [ 792.548339] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Creating folder: Instances. Parent ref: group-v259786. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.548818] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22fd0a5b-0eff-4db9-9f2b-8fdee4f01828 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.561151] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Created folder: Instances in parent group-v259786. [ 792.561483] env[62368]: DEBUG oslo.service.loopingcall [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.561848] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 792.562554] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12951a84-8b98-4c33-9ead-8ce607f7eef1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.592696] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.592696] env[62368]: value = "task-1198412" [ 792.592696] env[62368]: _type = "Task" [ 792.592696] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.604556] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198412, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.700186] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198408, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.717053] env[62368]: DEBUG nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 792.831119] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198409, 'name': ReconfigVM_Task, 'duration_secs': 0.376064} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.831516] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 3f1af54b-392f-432a-9ffa-a133da428f94/3f1af54b-392f-432a-9ffa-a133da428f94.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.832271] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca206807-3836-4e2b-b5e8-68fbcaa6a05c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.845355] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 792.845355] env[62368]: value = "task-1198413" [ 792.845355] env[62368]: _type = "Task" [ 792.845355] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.857957] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198413, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.872706] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "0f1bff84-1fb9-471a-b685-7d527b376b0a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.872978] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "0f1bff84-1fb9-471a-b685-7d527b376b0a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.873218] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "0f1bff84-1fb9-471a-b685-7d527b376b0a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.873406] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "0f1bff84-1fb9-471a-b685-7d527b376b0a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.873652] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "0f1bff84-1fb9-471a-b685-7d527b376b0a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.876829] env[62368]: INFO nova.compute.manager [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Terminating instance [ 792.878697] env[62368]: DEBUG nova.compute.manager [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.878862] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 792.879778] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc96206d-a689-4083-8edb-7888d27b7264 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.889345] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 792.891032] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49505ca5-084e-49eb-86f1-22404ec40885 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.905403] env[62368]: DEBUG oslo_vmware.api [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 792.905403] env[62368]: value = "task-1198414" [ 792.905403] env[62368]: _type = "Task" [ 792.905403] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.918660] env[62368]: DEBUG oslo_vmware.api [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.952229] env[62368]: DEBUG nova.network.neutron [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Successfully created port: bf67cd0c-ff81-451a-8e0e-6ee50bb77f78 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.969923] env[62368]: DEBUG nova.compute.manager [req-e531a764-1889-44da-a161-f544596fe9ff req-234419a4-33db-4ddf-8148-90222c78a396 service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Received event network-changed-444a47e3-2f56-4895-8a8b-7e5605dbfaf1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.972690] env[62368]: DEBUG nova.compute.manager [req-e531a764-1889-44da-a161-f544596fe9ff req-234419a4-33db-4ddf-8148-90222c78a396 service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Refreshing instance network info cache due to event network-changed-444a47e3-2f56-4895-8a8b-7e5605dbfaf1. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 792.972974] env[62368]: DEBUG oslo_concurrency.lockutils [req-e531a764-1889-44da-a161-f544596fe9ff req-234419a4-33db-4ddf-8148-90222c78a396 service nova] Acquiring lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.973146] env[62368]: DEBUG oslo_concurrency.lockutils [req-e531a764-1889-44da-a161-f544596fe9ff req-234419a4-33db-4ddf-8148-90222c78a396 service nova] Acquired lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.973349] env[62368]: DEBUG nova.network.neutron [req-e531a764-1889-44da-a161-f544596fe9ff req-234419a4-33db-4ddf-8148-90222c78a396 service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Refreshing network info cache for port 444a47e3-2f56-4895-8a8b-7e5605dbfaf1 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 793.116580] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198412, 'name': CreateVM_Task, 'duration_secs': 0.447161} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.116788] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 793.117667] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.118327] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.118327] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.121708] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb7e3255-edbd-4656-8d99-a1c2518daf7c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.127119] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 793.127119] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52496f2b-92bb-8ab1-4654-a3155f37df73" [ 793.127119] env[62368]: _type = "Task" [ 793.127119] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.148477] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52496f2b-92bb-8ab1-4654-a3155f37df73, 'name': SearchDatastore_Task, 'duration_secs': 0.01535} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.148838] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.149293] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.149568] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.149706] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.149875] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.153063] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1da8900-6bfe-4f21-a567-a94bedab61d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.166164] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.166390] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 793.167194] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e94b9431-0e9d-419a-b7c6-cbc7dc478bef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.176103] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 793.176103] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52664584-306a-d3ee-ecca-fe40b8d1c5b4" [ 793.176103] env[62368]: _type = "Task" [ 793.176103] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.183350] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8f3320-eb38-4733-8fb9-20f41b3a6de3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.188864] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52664584-306a-d3ee-ecca-fe40b8d1c5b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.196019] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf2713e-dec8-4346-abf7-2632f7b6adcb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.202318] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.671898} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.202960] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 87dc392f-7bba-44bb-ba38-2fd4e1c6129e/87dc392f-7bba-44bb-ba38-2fd4e1c6129e.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 793.203238] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.203501] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad803106-ea84-4449-897b-1de31eee5842 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.235463] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672ff4e0-f26a-4849-b89a-dc5a03b2ca76 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.241687] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 793.241687] env[62368]: value = "task-1198415" [ 793.241687] env[62368]: _type = "Task" [ 793.241687] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.248614] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ceaf82-f34f-48dd-a87f-9d1edc13a680 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.266883] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198415, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.276550] env[62368]: DEBUG nova.compute.provider_tree [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.355807] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198413, 'name': Rename_Task, 'duration_secs': 0.18715} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.356126] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 793.356749] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c3255f5-9495-404a-9dd2-b8ac34bd7cdb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.363880] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 793.363880] env[62368]: value = "task-1198416" [ 793.363880] env[62368]: _type = "Task" [ 793.363880] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.373783] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.414638] env[62368]: DEBUG oslo_vmware.api [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198414, 'name': PowerOffVM_Task, 'duration_secs': 0.37224} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.415264] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 793.415435] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 793.416713] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b25c9ad-5508-4436-9168-319fa320e8fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.501181] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 793.501430] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 793.501618] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Deleting the datastore file [datastore1] 0f1bff84-1fb9-471a-b685-7d527b376b0a {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.501907] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e10c362-1e1f-4af2-9844-b7a1b0fa03d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.510630] env[62368]: DEBUG oslo_vmware.api [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for the task: (returnval){ [ 793.510630] env[62368]: value = "task-1198418" [ 793.510630] env[62368]: _type = "Task" [ 793.510630] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.521686] env[62368]: DEBUG oslo_vmware.api [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198418, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.689570] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52664584-306a-d3ee-ecca-fe40b8d1c5b4, 'name': SearchDatastore_Task, 'duration_secs': 0.029145} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.690448] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f8cd125-eccf-418e-b993-d9d3c5b21680 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.697011] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 793.697011] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526f735c-5509-0f63-7c3c-64e63a31d61c" [ 793.697011] env[62368]: _type = "Task" [ 793.697011] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.705774] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526f735c-5509-0f63-7c3c-64e63a31d61c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.735328] env[62368]: DEBUG nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.755667] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198415, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102531} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.755970] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.757212] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93919f39-49ab-44f8-b8b1-6b7f311b875e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.785475] env[62368]: DEBUG nova.scheduler.client.report [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.802404] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 87dc392f-7bba-44bb-ba38-2fd4e1c6129e/87dc392f-7bba-44bb-ba38-2fd4e1c6129e.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.806480] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.806785] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.806978] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.807329] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.807483] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.807650] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.808142] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.808483] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.808705] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.808885] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.809229] env[62368]: DEBUG nova.virt.hardware [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.809916] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f570b80a-7297-47fb-a0fb-9d03d786aa41 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.827584] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55d3545-7adf-47ed-9b47-fdcdf8ae5ddc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.831543] env[62368]: DEBUG nova.network.neutron [req-e531a764-1889-44da-a161-f544596fe9ff req-234419a4-33db-4ddf-8148-90222c78a396 service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updated VIF entry in instance network info cache for port 444a47e3-2f56-4895-8a8b-7e5605dbfaf1. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 793.831948] env[62368]: DEBUG nova.network.neutron [req-e531a764-1889-44da-a161-f544596fe9ff req-234419a4-33db-4ddf-8148-90222c78a396 service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance_info_cache with network_info: [{"id": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "address": "fa:16:3e:98:a0:d2", "network": {"id": "34f156be-cab3-45a2-8b52-74dd4cd1e42d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2a054e975862463bb842479ad19393ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap444a47e3-2f", "ovs_interfaceid": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.840859] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b78c5b-c4db-4b53-81a2-93526ca15468 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.847841] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 793.847841] env[62368]: value = "task-1198419" [ 793.847841] env[62368]: _type = "Task" [ 793.847841] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.867215] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198419, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.877989] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198416, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.025939] env[62368]: DEBUG oslo_vmware.api [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Task: {'id': task-1198418, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.499332} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.025939] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.025939] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 794.026159] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 794.026271] env[62368]: INFO nova.compute.manager [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 794.026536] env[62368]: DEBUG oslo.service.loopingcall [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.026716] env[62368]: DEBUG nova.compute.manager [-] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.026808] env[62368]: DEBUG nova.network.neutron [-] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 794.208942] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526f735c-5509-0f63-7c3c-64e63a31d61c, 'name': SearchDatastore_Task, 'duration_secs': 0.024855} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.209247] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.209508] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071/a0cea538-b162-4504-ac34-803a2d5a8071.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 794.209779] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54abfe7c-44c3-42db-8d7f-752c6974d32f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.218408] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 794.218408] env[62368]: value = "task-1198420" [ 794.218408] env[62368]: _type = "Task" [ 794.218408] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.227141] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198420, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.304366] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.305552] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.805s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.305789] env[62368]: DEBUG nova.objects.instance [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'resources' on Instance uuid af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 794.337170] env[62368]: DEBUG oslo_concurrency.lockutils [req-e531a764-1889-44da-a161-f544596fe9ff req-234419a4-33db-4ddf-8148-90222c78a396 service nova] Releasing lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.358580] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198419, 'name': ReconfigVM_Task, 'duration_secs': 0.409769} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.358858] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 87dc392f-7bba-44bb-ba38-2fd4e1c6129e/87dc392f-7bba-44bb-ba38-2fd4e1c6129e.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.359498] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0008c40-df2f-4c7d-a430-bdb956363d00 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.366626] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 794.366626] env[62368]: value = "task-1198421" [ 794.366626] env[62368]: _type = "Task" [ 794.366626] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.377539] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198421, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.380442] env[62368]: DEBUG oslo_vmware.api [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198416, 'name': PowerOnVM_Task, 'duration_secs': 0.519237} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.380778] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 794.380965] env[62368]: DEBUG nova.compute.manager [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.381761] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc48c6c8-0358-458e-a17f-1c62226461d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.513521] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "2a46c954-449e-4d62-be80-add1040ed4c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.513834] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "2a46c954-449e-4d62-be80-add1040ed4c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.732357] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198420, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.808887] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "a793660e-3ee3-43b7-ae0a-76a2422bc27b" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.809733] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "a793660e-3ee3-43b7-ae0a-76a2422bc27b" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.879894] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198421, 'name': Rename_Task, 'duration_secs': 0.149623} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.880223] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 794.880539] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74773128-58ed-4d72-aa12-c0fdd364d63a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.889755] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 794.889755] env[62368]: value = "task-1198422" [ 794.889755] env[62368]: _type = "Task" [ 794.889755] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.902951] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198422, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.905150] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.906462] env[62368]: DEBUG nova.compute.manager [req-fc4434ac-6419-4dfd-8c70-d844e5b268e2 req-bfc711e9-b7e3-4259-972f-de0ca306b2ee service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Received event network-vif-plugged-bf67cd0c-ff81-451a-8e0e-6ee50bb77f78 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.906614] env[62368]: DEBUG oslo_concurrency.lockutils [req-fc4434ac-6419-4dfd-8c70-d844e5b268e2 req-bfc711e9-b7e3-4259-972f-de0ca306b2ee service nova] Acquiring lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.906824] env[62368]: DEBUG oslo_concurrency.lockutils [req-fc4434ac-6419-4dfd-8c70-d844e5b268e2 req-bfc711e9-b7e3-4259-972f-de0ca306b2ee service nova] Lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.906995] env[62368]: DEBUG oslo_concurrency.lockutils [req-fc4434ac-6419-4dfd-8c70-d844e5b268e2 req-bfc711e9-b7e3-4259-972f-de0ca306b2ee service nova] Lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.907189] env[62368]: DEBUG nova.compute.manager [req-fc4434ac-6419-4dfd-8c70-d844e5b268e2 req-bfc711e9-b7e3-4259-972f-de0ca306b2ee service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] No waiting events found dispatching network-vif-plugged-bf67cd0c-ff81-451a-8e0e-6ee50bb77f78 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 794.907360] env[62368]: WARNING nova.compute.manager [req-fc4434ac-6419-4dfd-8c70-d844e5b268e2 req-bfc711e9-b7e3-4259-972f-de0ca306b2ee service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Received unexpected event network-vif-plugged-bf67cd0c-ff81-451a-8e0e-6ee50bb77f78 for instance with vm_state building and task_state spawning. [ 794.991338] env[62368]: DEBUG nova.network.neutron [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Successfully updated port: bf67cd0c-ff81-451a-8e0e-6ee50bb77f78 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 795.189947] env[62368]: DEBUG nova.compute.manager [req-7a98227a-b689-41af-aa2a-910cce3a4dcb req-967acc17-4dec-427c-814a-a19b6a66bd15 service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Received event network-vif-deleted-e7892749-2fef-4fa0-ba5b-348c58477c93 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.190180] env[62368]: INFO nova.compute.manager [req-7a98227a-b689-41af-aa2a-910cce3a4dcb req-967acc17-4dec-427c-814a-a19b6a66bd15 service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Neutron deleted interface e7892749-2fef-4fa0-ba5b-348c58477c93; detaching it from the instance and deleting it from the info cache [ 795.190423] env[62368]: DEBUG nova.network.neutron [req-7a98227a-b689-41af-aa2a-910cce3a4dcb req-967acc17-4dec-427c-814a-a19b6a66bd15 service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.207793] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89bec09-58f4-4fcc-a336-44a3cb76d02e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.220021] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d304d067-eb0f-4b46-be96-11b249320119 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.258732] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198420, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.260642] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf96d736-8323-4c2c-a17d-30eda7ce6096 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.270503] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75acd8a4-eb8f-4dca-8e1d-fa8e5cdfeffa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.286400] env[62368]: DEBUG nova.compute.provider_tree [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.299471] env[62368]: DEBUG nova.network.neutron [-] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.319893] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "a793660e-3ee3-43b7-ae0a-76a2422bc27b" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.508s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.319893] env[62368]: DEBUG nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.401190] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198422, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.494254] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.494441] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.494546] env[62368]: DEBUG nova.network.neutron [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 795.694228] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76412508-65d9-4f58-8e0c-e24cf470583a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.707603] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0ad822-65e9-4ed6-a368-64fd9662b13e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.739188] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198420, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.750098] env[62368]: DEBUG nova.compute.manager [req-7a98227a-b689-41af-aa2a-910cce3a4dcb req-967acc17-4dec-427c-814a-a19b6a66bd15 service nova] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Detach interface failed, port_id=e7892749-2fef-4fa0-ba5b-348c58477c93, reason: Instance 0f1bff84-1fb9-471a-b685-7d527b376b0a could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 795.790277] env[62368]: DEBUG nova.scheduler.client.report [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.802596] env[62368]: INFO nova.compute.manager [-] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Took 1.78 seconds to deallocate network for instance. [ 795.823408] env[62368]: DEBUG nova.compute.utils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 795.825110] env[62368]: DEBUG nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 795.825460] env[62368]: DEBUG nova.network.neutron [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 795.891224] env[62368]: DEBUG nova.policy [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d297a4b095645b885a408e7d8cd7176', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '898bdcb117554186b1fa93017eab0270', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 795.903973] env[62368]: DEBUG oslo_vmware.api [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198422, 'name': PowerOnVM_Task, 'duration_secs': 0.565547} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.904240] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 795.904452] env[62368]: INFO nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Took 9.26 seconds to spawn the instance on the hypervisor. [ 795.904637] env[62368]: DEBUG nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.905809] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20c365f-859d-421b-ac15-29d1fd9eca36 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.949029] env[62368]: DEBUG nova.compute.manager [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.949029] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0aee8ed-6fbf-4a76-892d-e2e0e8834f77 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.074456] env[62368]: DEBUG nova.network.neutron [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 796.243232] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198420, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.715455} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.243518] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071/a0cea538-b162-4504-ac34-803a2d5a8071.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 796.243733] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.244742] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad883130-96ed-4146-ac8d-5751d6103bec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.257444] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 796.257444] env[62368]: value = "task-1198423" [ 796.257444] env[62368]: _type = "Task" [ 796.257444] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.265293] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.303686] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.998s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.306488] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.521s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.306757] env[62368]: DEBUG nova.objects.instance [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lazy-loading 'resources' on Instance uuid 4fe7b1a2-f894-4131-9456-fb4df8a4532d {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.309130] env[62368]: DEBUG nova.network.neutron [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Successfully created port: 31693720-039f-4059-b582-9c57cbe54814 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.315178] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.328873] env[62368]: DEBUG nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.344659] env[62368]: INFO nova.scheduler.client.report [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted allocations for instance af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03 [ 796.368645] env[62368]: DEBUG nova.network.neutron [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Updating instance_info_cache with network_info: [{"id": "bf67cd0c-ff81-451a-8e0e-6ee50bb77f78", "address": "fa:16:3e:90:7b:bf", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf67cd0c-ff", "ovs_interfaceid": "bf67cd0c-ff81-451a-8e0e-6ee50bb77f78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.428699] env[62368]: INFO nova.compute.manager [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Took 37.65 seconds to build instance. [ 796.461629] env[62368]: INFO nova.compute.manager [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] instance snapshotting [ 796.461629] env[62368]: DEBUG nova.objects.instance [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'flavor' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.576912] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "3f1af54b-392f-432a-9ffa-a133da428f94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.576912] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "3f1af54b-392f-432a-9ffa-a133da428f94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.576912] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "3f1af54b-392f-432a-9ffa-a133da428f94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.576912] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "3f1af54b-392f-432a-9ffa-a133da428f94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.576912] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "3f1af54b-392f-432a-9ffa-a133da428f94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.580028] env[62368]: INFO nova.compute.manager [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Terminating instance [ 796.585015] env[62368]: DEBUG nova.compute.manager [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 796.585015] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 796.585015] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbe1da0-528b-461a-89e8-07fa3b8036f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.594411] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 796.597034] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aadb2de7-f8e1-4ee1-994d-35f148d329d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.608965] env[62368]: DEBUG oslo_vmware.api [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 796.608965] env[62368]: value = "task-1198424" [ 796.608965] env[62368]: _type = "Task" [ 796.608965] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.621332] env[62368]: DEBUG oslo_vmware.api [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198424, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.771098] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198423, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.195098} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.771517] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.774395] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f747a5d8-ab72-4cbe-aece-3b89b5e6192c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.803081] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071/a0cea538-b162-4504-ac34-803a2d5a8071.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.804043] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd0f0cdc-f95f-493f-b39b-98ed86e67d7f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.828862] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 796.828862] env[62368]: value = "task-1198425" [ 796.828862] env[62368]: _type = "Task" [ 796.828862] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.841906] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198425, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.854430] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e2b0b43f-2b58-4ad2-9423-af470cfe1e59 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.923s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.870814] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.875014] env[62368]: DEBUG nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Instance network_info: |[{"id": "bf67cd0c-ff81-451a-8e0e-6ee50bb77f78", "address": "fa:16:3e:90:7b:bf", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf67cd0c-ff", "ovs_interfaceid": "bf67cd0c-ff81-451a-8e0e-6ee50bb77f78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 796.875424] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:7b:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8f40f5c4-c146-449c-884d-6f884dcf2acf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf67cd0c-ff81-451a-8e0e-6ee50bb77f78', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.886757] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating folder: Project (026a3176abc34001ab3506445226777f). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 796.891141] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dcd52b27-f276-46ce-b5b0-20ec0e13e15e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.908751] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created folder: Project (026a3176abc34001ab3506445226777f) in parent group-v259706. [ 796.908983] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating folder: Instances. Parent ref: group-v259789. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 796.909791] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-91202aa3-6b2d-46da-bd4a-db07852d8515 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.921286] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created folder: Instances in parent group-v259789. [ 796.921809] env[62368]: DEBUG oslo.service.loopingcall [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.922240] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 796.922685] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8843067-f697-4094-89be-f29f3eea8c6e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.943255] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6a48e81a-fe1f-48b7-8257-7b4fe2013d2b tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.801s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.950245] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.950245] env[62368]: value = "task-1198428" [ 796.950245] env[62368]: _type = "Task" [ 796.950245] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.968859] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198428, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.970507] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5725d64-906f-4040-94d9-82d4370e0118 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.001998] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89caabb5-d900-42a4-b1a4-0f8d6438f1a4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.065281] env[62368]: DEBUG nova.compute.manager [req-49c12f31-f615-4f14-9124-611a346275fe req-a37b306d-7625-4d48-a6ec-8e00f675f6a2 service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Received event network-changed-bf67cd0c-ff81-451a-8e0e-6ee50bb77f78 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.065490] env[62368]: DEBUG nova.compute.manager [req-49c12f31-f615-4f14-9124-611a346275fe req-a37b306d-7625-4d48-a6ec-8e00f675f6a2 service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Refreshing instance network info cache due to event network-changed-bf67cd0c-ff81-451a-8e0e-6ee50bb77f78. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 797.065718] env[62368]: DEBUG oslo_concurrency.lockutils [req-49c12f31-f615-4f14-9124-611a346275fe req-a37b306d-7625-4d48-a6ec-8e00f675f6a2 service nova] Acquiring lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.065868] env[62368]: DEBUG oslo_concurrency.lockutils [req-49c12f31-f615-4f14-9124-611a346275fe req-a37b306d-7625-4d48-a6ec-8e00f675f6a2 service nova] Acquired lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.066599] env[62368]: DEBUG nova.network.neutron [req-49c12f31-f615-4f14-9124-611a346275fe req-a37b306d-7625-4d48-a6ec-8e00f675f6a2 service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Refreshing network info cache for port bf67cd0c-ff81-451a-8e0e-6ee50bb77f78 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 797.124286] env[62368]: DEBUG oslo_vmware.api [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198424, 'name': PowerOffVM_Task, 'duration_secs': 0.233301} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.126097] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 797.126294] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 797.133219] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c97e33d-62c1-427c-b9eb-0491c2382d2d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.262630] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88306db3-d848-4eb2-9c1b-d0b7adecb8c5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.275077] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f194e5fd-60c7-4e7a-b9a1-146e981bea9a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.314228] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b23a5d7-9218-4013-8c6a-4e42dc34139d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.316922] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 797.317160] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 797.317307] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleting the datastore file [datastore2] 3f1af54b-392f-432a-9ffa-a133da428f94 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.318310] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4310bf2f-d7c2-4442-82fb-3f0553840483 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.327161] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cef617-7765-4598-bd8e-2adaf7a7d538 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.331321] env[62368]: DEBUG oslo_vmware.api [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 797.331321] env[62368]: value = "task-1198430" [ 797.331321] env[62368]: _type = "Task" [ 797.331321] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.348417] env[62368]: DEBUG nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.349793] env[62368]: DEBUG nova.compute.provider_tree [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.362265] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "58efc1cf-2469-41be-bf99-fc7b8c72113c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.362517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "58efc1cf-2469-41be-bf99-fc7b8c72113c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.362738] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198425, 'name': ReconfigVM_Task, 'duration_secs': 0.330019} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.362942] env[62368]: DEBUG oslo_vmware.api [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198430, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.364138] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfigured VM instance instance-0000003d to attach disk [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071/a0cea538-b162-4504-ac34-803a2d5a8071.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.364847] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b12ab97-105f-4575-b3eb-b98b4855ca55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.375128] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.375383] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.375544] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.375727] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.375878] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.376066] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.376339] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.376507] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.376679] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.376858] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.377044] env[62368]: DEBUG nova.virt.hardware [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.378368] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7b1839-d788-45b8-a8b0-f3a1e43f33af {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.382965] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 797.382965] env[62368]: value = "task-1198431" [ 797.382965] env[62368]: _type = "Task" [ 797.382965] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.390324] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99adaf06-9114-424c-a2a3-75e567993c86 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.398927] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198431, 'name': Rename_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.445877] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.464448] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198428, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.514633] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 797.515076] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8c46ea19-0624-4960-bb9b-aab559424a12 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.525009] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 797.525009] env[62368]: value = "task-1198432" [ 797.525009] env[62368]: _type = "Task" [ 797.525009] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.534079] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198432, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.771374] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.771819] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.771819] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.771983] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.772195] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.775242] env[62368]: INFO nova.compute.manager [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Terminating instance [ 797.779231] env[62368]: DEBUG nova.compute.manager [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 797.779560] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 797.780841] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0961e7ac-c87d-4cba-b81f-44b1446c18d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.793812] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 797.794272] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b95d1eb-9c8b-42bb-a2f4-4dd0197b5f98 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.806717] env[62368]: DEBUG oslo_vmware.api [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 797.806717] env[62368]: value = "task-1198433" [ 797.806717] env[62368]: _type = "Task" [ 797.806717] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.818599] env[62368]: DEBUG oslo_vmware.api [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198433, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.843342] env[62368]: DEBUG oslo_vmware.api [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198430, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254868} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.844750] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.844750] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 797.844750] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 797.844750] env[62368]: INFO nova.compute.manager [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Took 1.26 seconds to destroy the instance on the hypervisor. [ 797.844750] env[62368]: DEBUG oslo.service.loopingcall [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.844750] env[62368]: DEBUG nova.compute.manager [-] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.844750] env[62368]: DEBUG nova.network.neutron [-] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.857098] env[62368]: DEBUG nova.scheduler.client.report [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.903605] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198431, 'name': Rename_Task, 'duration_secs': 0.262272} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.903605] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 797.904715] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1c16ec4-8132-4959-ae60-12c3dfeee4e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.915330] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 797.915330] env[62368]: value = "task-1198434" [ 797.915330] env[62368]: _type = "Task" [ 797.915330] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.927124] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.969806] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198428, 'name': CreateVM_Task, 'duration_secs': 0.631345} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.969936] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 797.970679] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.970891] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.971273] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.972521] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-644dc774-8ebf-42ed-b5b3-90e726607955 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.977981] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 797.977981] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52560406-2da0-a8b7-63ff-60385e899385" [ 797.977981] env[62368]: _type = "Task" [ 797.977981] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.983017] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.997052] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52560406-2da0-a8b7-63ff-60385e899385, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.036579] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198432, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.071816] env[62368]: DEBUG nova.network.neutron [req-49c12f31-f615-4f14-9124-611a346275fe req-a37b306d-7625-4d48-a6ec-8e00f675f6a2 service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Updated VIF entry in instance network info cache for port bf67cd0c-ff81-451a-8e0e-6ee50bb77f78. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 798.072569] env[62368]: DEBUG nova.network.neutron [req-49c12f31-f615-4f14-9124-611a346275fe req-a37b306d-7625-4d48-a6ec-8e00f675f6a2 service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Updating instance_info_cache with network_info: [{"id": "bf67cd0c-ff81-451a-8e0e-6ee50bb77f78", "address": "fa:16:3e:90:7b:bf", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf67cd0c-ff", "ovs_interfaceid": "bf67cd0c-ff81-451a-8e0e-6ee50bb77f78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.231174] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.231174] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.231174] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.231174] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.231364] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.235133] env[62368]: INFO nova.compute.manager [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Terminating instance [ 798.237644] env[62368]: DEBUG nova.compute.manager [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 798.237885] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 798.239071] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3385633-940d-433c-a9e9-a988ed9c93fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.249319] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 798.250062] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fefec26-fd96-44f5-bcd2-d575222f2142 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.260605] env[62368]: DEBUG oslo_vmware.api [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 798.260605] env[62368]: value = "task-1198435" [ 798.260605] env[62368]: _type = "Task" [ 798.260605] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.272844] env[62368]: DEBUG oslo_vmware.api [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.320054] env[62368]: DEBUG oslo_vmware.api [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198433, 'name': PowerOffVM_Task, 'duration_secs': 0.269009} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.320054] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 798.320054] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 798.320435] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3a6f61f-b42b-4724-a478-80acf9ed35fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.362884] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.056s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.367024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.636s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.369376] env[62368]: INFO nova.compute.claims [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.378088] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.378297] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.392210] env[62368]: INFO nova.scheduler.client.report [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Deleted allocations for instance 4fe7b1a2-f894-4131-9456-fb4df8a4532d [ 798.405583] env[62368]: DEBUG nova.network.neutron [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Successfully updated port: 31693720-039f-4059-b582-9c57cbe54814 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.434269] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198434, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.496733] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52560406-2da0-a8b7-63ff-60385e899385, 'name': SearchDatastore_Task, 'duration_secs': 0.012563} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.498308] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.498604] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.498866] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.499703] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.499703] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.499703] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 798.499851] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 798.500048] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Deleting the datastore file [datastore1] 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.500328] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ef087a5-d5a2-483e-9a42-9760c16e9edc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.503047] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee398935-2700-4711-a6e3-75b8f62f985f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.511371] env[62368]: DEBUG oslo_vmware.api [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 798.511371] env[62368]: value = "task-1198437" [ 798.511371] env[62368]: _type = "Task" [ 798.511371] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.516759] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.516966] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 798.518468] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2081f6b2-7496-4790-847e-1a4464d17f6e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.524936] env[62368]: DEBUG oslo_vmware.api [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.531667] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 798.531667] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52dc3494-ad5f-7df2-9f7d-dc2a8d630320" [ 798.531667] env[62368]: _type = "Task" [ 798.531667] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.539035] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198432, 'name': CreateSnapshot_Task, 'duration_secs': 0.912526} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.539728] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 798.540511] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce3071d-756e-45e3-a5c5-4fdb425a5e9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.547663] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52dc3494-ad5f-7df2-9f7d-dc2a8d630320, 'name': SearchDatastore_Task, 'duration_secs': 0.014815} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.547663] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c72c5fb8-a4e9-4976-9976-84170a101bad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.561129] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 798.561129] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f6ae29-db69-8d8c-8974-6e90bfa0e8bd" [ 798.561129] env[62368]: _type = "Task" [ 798.561129] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.571906] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f6ae29-db69-8d8c-8974-6e90bfa0e8bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.575600] env[62368]: DEBUG oslo_concurrency.lockutils [req-49c12f31-f615-4f14-9124-611a346275fe req-a37b306d-7625-4d48-a6ec-8e00f675f6a2 service nova] Releasing lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.771887] env[62368]: DEBUG oslo_vmware.api [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198435, 'name': PowerOffVM_Task, 'duration_secs': 0.353335} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.772734] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 798.775177] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 798.775177] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-024af180-d485-4d6a-8b36-3ace77a28206 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.866851] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 798.866851] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 798.866851] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Deleting the datastore file [datastore2] 87dc392f-7bba-44bb-ba38-2fd4e1c6129e {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.866851] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-73f48035-0411-4d07-a03a-13a686a2cece {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.876017] env[62368]: DEBUG oslo_vmware.api [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for the task: (returnval){ [ 798.876017] env[62368]: value = "task-1198439" [ 798.876017] env[62368]: _type = "Task" [ 798.876017] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.888306] env[62368]: DEBUG oslo_vmware.api [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198439, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.891355] env[62368]: DEBUG nova.network.neutron [-] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.906511] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e283e665-fa21-427e-a27a-4e0af57b93b3 tempest-ServerRescueTestJSONUnderV235-164048778 tempest-ServerRescueTestJSONUnderV235-164048778-project-member] Lock "4fe7b1a2-f894-4131-9456-fb4df8a4532d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.489s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.912845] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "refresh_cache-54083854-e314-44a7-b4b0-fbef5fa6b1ef" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.912845] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquired lock "refresh_cache-54083854-e314-44a7-b4b0-fbef5fa6b1ef" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.912845] env[62368]: DEBUG nova.network.neutron [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 798.928489] env[62368]: DEBUG oslo_vmware.api [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198434, 'name': PowerOnVM_Task, 'duration_secs': 0.62604} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.929134] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 798.929553] env[62368]: INFO nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Took 9.80 seconds to spawn the instance on the hypervisor. [ 798.930037] env[62368]: DEBUG nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 798.932504] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428e9d00-b8ad-436e-9a7d-8f3dac964d15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.026223] env[62368]: DEBUG oslo_vmware.api [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230776} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.026698] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.027073] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 799.027413] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 799.027735] env[62368]: INFO nova.compute.manager [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Took 1.25 seconds to destroy the instance on the hypervisor. [ 799.028535] env[62368]: DEBUG oslo.service.loopingcall [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.028916] env[62368]: DEBUG nova.compute.manager [-] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.029195] env[62368]: DEBUG nova.network.neutron [-] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 799.071763] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 799.072175] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0d0ad83e-4dd6-4511-93a1-698dc999ef6a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.086847] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f6ae29-db69-8d8c-8974-6e90bfa0e8bd, 'name': SearchDatastore_Task, 'duration_secs': 0.020761} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.088544] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.088821] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b79f0e79-9e3a-47c7-9949-8743601ec6c5/b79f0e79-9e3a-47c7-9949-8743601ec6c5.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 799.089186] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 799.089186] env[62368]: value = "task-1198440" [ 799.089186] env[62368]: _type = "Task" [ 799.089186] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.092022] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-00cb7a6d-7c16-4419-bffa-42eb93b6b715 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.101736] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198440, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.103247] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 799.103247] env[62368]: value = "task-1198441" [ 799.103247] env[62368]: _type = "Task" [ 799.103247] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.114194] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198441, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.164089] env[62368]: DEBUG nova.compute.manager [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Received event network-vif-plugged-31693720-039f-4059-b582-9c57cbe54814 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.164340] env[62368]: DEBUG oslo_concurrency.lockutils [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] Acquiring lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.165018] env[62368]: DEBUG oslo_concurrency.lockutils [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] Lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.165263] env[62368]: DEBUG oslo_concurrency.lockutils [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] Lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.165457] env[62368]: DEBUG nova.compute.manager [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] No waiting events found dispatching network-vif-plugged-31693720-039f-4059-b582-9c57cbe54814 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.165625] env[62368]: WARNING nova.compute.manager [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Received unexpected event network-vif-plugged-31693720-039f-4059-b582-9c57cbe54814 for instance with vm_state building and task_state spawning. [ 799.165815] env[62368]: DEBUG nova.compute.manager [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Received event network-vif-deleted-c96365aa-f01b-493f-b516-a4e22b5b43e1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.166009] env[62368]: DEBUG nova.compute.manager [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Received event network-changed-31693720-039f-4059-b582-9c57cbe54814 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.166210] env[62368]: DEBUG nova.compute.manager [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Refreshing instance network info cache due to event network-changed-31693720-039f-4059-b582-9c57cbe54814. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.166602] env[62368]: DEBUG oslo_concurrency.lockutils [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] Acquiring lock "refresh_cache-54083854-e314-44a7-b4b0-fbef5fa6b1ef" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.393219] env[62368]: DEBUG oslo_vmware.api [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Task: {'id': task-1198439, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187714} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.394978] env[62368]: INFO nova.compute.manager [-] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Took 1.55 seconds to deallocate network for instance. [ 799.395891] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.396504] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 799.397198] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 799.397516] env[62368]: INFO nova.compute.manager [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 799.397725] env[62368]: DEBUG oslo.service.loopingcall [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.402595] env[62368]: DEBUG nova.compute.manager [-] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.402720] env[62368]: DEBUG nova.network.neutron [-] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 799.464172] env[62368]: INFO nova.compute.manager [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Took 40.15 seconds to build instance. [ 799.468868] env[62368]: DEBUG nova.network.neutron [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.608105] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198440, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.621330] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198441, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.810587] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebbc36b-5041-4e6a-9b83-b486f25baf73 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.820355] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bdd4bd-88b3-4f04-abf0-45f2de5f5562 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.859863] env[62368]: DEBUG nova.network.neutron [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Updating instance_info_cache with network_info: [{"id": "31693720-039f-4059-b582-9c57cbe54814", "address": "fa:16:3e:48:49:ba", "network": {"id": "cacf6e9a-7be1-492f-81e7-a38f37a2ac41", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1409244324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "898bdcb117554186b1fa93017eab0270", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31693720-03", "ovs_interfaceid": "31693720-039f-4059-b582-9c57cbe54814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.861948] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a83ff21-ea8c-4a7f-b7bd-17380d68ffeb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.870908] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b43dff-670c-494c-be82-00f9bfefffa8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.886865] env[62368]: DEBUG nova.compute.provider_tree [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.906202] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.939186] env[62368]: DEBUG nova.network.neutron [-] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.969616] env[62368]: DEBUG oslo_concurrency.lockutils [None req-90d59b27-13ff-4dcd-9567-3942aa34ba8b tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "a0cea538-b162-4504-ac34-803a2d5a8071" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.658s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.106244] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198440, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.118994] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198441, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554452} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.118994] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b79f0e79-9e3a-47c7-9949-8743601ec6c5/b79f0e79-9e3a-47c7-9949-8743601ec6c5.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 800.118994] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 800.119133] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55145916-2de4-4315-85c5-ddb89cf3865c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.127472] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 800.127472] env[62368]: value = "task-1198442" [ 800.127472] env[62368]: _type = "Task" [ 800.127472] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.140163] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.365715] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Releasing lock "refresh_cache-54083854-e314-44a7-b4b0-fbef5fa6b1ef" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.365715] env[62368]: DEBUG nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Instance network_info: |[{"id": "31693720-039f-4059-b582-9c57cbe54814", "address": "fa:16:3e:48:49:ba", "network": {"id": "cacf6e9a-7be1-492f-81e7-a38f37a2ac41", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1409244324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "898bdcb117554186b1fa93017eab0270", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31693720-03", "ovs_interfaceid": "31693720-039f-4059-b582-9c57cbe54814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.366237] env[62368]: DEBUG oslo_concurrency.lockutils [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] Acquired lock "refresh_cache-54083854-e314-44a7-b4b0-fbef5fa6b1ef" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.366650] env[62368]: DEBUG nova.network.neutron [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Refreshing network info cache for port 31693720-039f-4059-b582-9c57cbe54814 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.368030] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:49:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbd7899c-c96e-47fc-9141-5803b646917a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31693720-039f-4059-b582-9c57cbe54814', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.376886] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Creating folder: Project (898bdcb117554186b1fa93017eab0270). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.379015] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-469c0564-db75-4bba-aede-3c6f7482ecc9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.389690] env[62368]: DEBUG nova.scheduler.client.report [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.398970] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Created folder: Project (898bdcb117554186b1fa93017eab0270) in parent group-v259706. [ 800.399177] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Creating folder: Instances. Parent ref: group-v259794. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.400164] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-beb519fe-7bb2-47f7-9c55-bf99288d4b28 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.415780] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Created folder: Instances in parent group-v259794. [ 800.418799] env[62368]: DEBUG oslo.service.loopingcall [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.418799] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 800.418799] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae599eef-df50-42a7-aba3-7572c6daffb4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.442829] env[62368]: INFO nova.compute.manager [-] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Took 1.41 seconds to deallocate network for instance. [ 800.443355] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.443355] env[62368]: value = "task-1198445" [ 800.443355] env[62368]: _type = "Task" [ 800.443355] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.457771] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198445, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.470920] env[62368]: DEBUG nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.607375] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198440, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.627829] env[62368]: DEBUG nova.network.neutron [-] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.644689] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075798} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.645023] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.645847] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7e8822-a87a-4287-b31c-daa66c9715ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.671175] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] b79f0e79-9e3a-47c7-9949-8743601ec6c5/b79f0e79-9e3a-47c7-9949-8743601ec6c5.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.671867] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b975ffc9-9a0f-4a48-9d11-44bbf0de4507 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.693450] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 800.693450] env[62368]: value = "task-1198446" [ 800.693450] env[62368]: _type = "Task" [ 800.693450] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.703595] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198446, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.900949] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.901565] env[62368]: DEBUG nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.904356] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.447s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.905967] env[62368]: INFO nova.compute.claims [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.958093] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.964762] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198445, 'name': CreateVM_Task, 'duration_secs': 0.426154} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.964875] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 800.965812] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.966038] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.966441] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 800.966719] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccffef29-de1a-4995-98f5-46916e1eb33a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.972976] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 800.972976] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52086149-e48c-c147-0d5c-4a49f754e774" [ 800.972976] env[62368]: _type = "Task" [ 800.972976] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.991316] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52086149-e48c-c147-0d5c-4a49f754e774, 'name': SearchDatastore_Task, 'duration_secs': 0.011103} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.991316] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.991316] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 800.991316] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.991316] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.991316] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 800.991502] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-787c8cd5-15f0-46a5-bd6a-6302c3fe7314 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.004761] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.005014] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 801.005765] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c495897-2760-4cd7-a475-e11b5ea9f782 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.012563] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.014740] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 801.014740] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c2b696-5c9d-392d-cf3d-7e0ae39cd1dc" [ 801.014740] env[62368]: _type = "Task" [ 801.014740] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.025655] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c2b696-5c9d-392d-cf3d-7e0ae39cd1dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.110730] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198440, 'name': CloneVM_Task} progress is 95%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.135781] env[62368]: INFO nova.compute.manager [-] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Took 1.73 seconds to deallocate network for instance. [ 801.214417] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198446, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.403972] env[62368]: DEBUG nova.network.neutron [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Updated VIF entry in instance network info cache for port 31693720-039f-4059-b582-9c57cbe54814. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.404409] env[62368]: DEBUG nova.network.neutron [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Updating instance_info_cache with network_info: [{"id": "31693720-039f-4059-b582-9c57cbe54814", "address": "fa:16:3e:48:49:ba", "network": {"id": "cacf6e9a-7be1-492f-81e7-a38f37a2ac41", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1409244324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "898bdcb117554186b1fa93017eab0270", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbd7899c-c96e-47fc-9141-5803b646917a", "external-id": "nsx-vlan-transportzone-333", "segmentation_id": 333, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31693720-03", "ovs_interfaceid": "31693720-039f-4059-b582-9c57cbe54814", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.414401] env[62368]: DEBUG nova.compute.utils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.425043] env[62368]: DEBUG nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.425043] env[62368]: DEBUG nova.network.neutron [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 801.435570] env[62368]: DEBUG nova.compute.manager [req-22ff299d-c6ed-44ac-8227-fa6b15869e80 req-6f854aa3-2189-403a-b499-8a3e878d2b00 service nova] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Received event network-vif-deleted-9897441b-2e19-4f96-8124-6aae0b31e308 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.435570] env[62368]: DEBUG nova.compute.manager [req-22ff299d-c6ed-44ac-8227-fa6b15869e80 req-6f854aa3-2189-403a-b499-8a3e878d2b00 service nova] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Received event network-vif-deleted-4b9518d9-5b60-41b5-a631-9eeba2186936 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.527056] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c2b696-5c9d-392d-cf3d-7e0ae39cd1dc, 'name': SearchDatastore_Task, 'duration_secs': 0.011294} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.527056] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ecd4479-d75d-43fa-a443-4a16020e2841 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.539407] env[62368]: DEBUG nova.policy [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc2d082d49ba4dd480535c2c56501fff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e250606f3df240eb93ce45925b00ad43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.543831] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 801.543831] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527cfca3-5f4d-73bb-8b20-9ed7f52d198f" [ 801.543831] env[62368]: _type = "Task" [ 801.543831] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.552489] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527cfca3-5f4d-73bb-8b20-9ed7f52d198f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.612870] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198440, 'name': CloneVM_Task, 'duration_secs': 2.253316} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.613210] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Created linked-clone VM from snapshot [ 801.614053] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73a0936-2a78-4fb4-9dd5-7669210c3834 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.630211] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Uploading image 3b178833-ecf1-4a1f-8ffe-ee20324be047 {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 801.644197] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.667332] env[62368]: DEBUG oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 801.667332] env[62368]: value = "vm-259793" [ 801.667332] env[62368]: _type = "VirtualMachine" [ 801.667332] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 801.667704] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-561c7d7b-5852-415c-965c-2b40a348b7d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.680391] env[62368]: DEBUG oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease: (returnval){ [ 801.680391] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52482d53-557e-b050-9dd6-1c8210f9799f" [ 801.680391] env[62368]: _type = "HttpNfcLease" [ 801.680391] env[62368]: } obtained for exporting VM: (result){ [ 801.680391] env[62368]: value = "vm-259793" [ 801.680391] env[62368]: _type = "VirtualMachine" [ 801.680391] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 801.682524] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the lease: (returnval){ [ 801.682524] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52482d53-557e-b050-9dd6-1c8210f9799f" [ 801.682524] env[62368]: _type = "HttpNfcLease" [ 801.682524] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 801.693718] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 801.693718] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52482d53-557e-b050-9dd6-1c8210f9799f" [ 801.693718] env[62368]: _type = "HttpNfcLease" [ 801.693718] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 801.707649] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198446, 'name': ReconfigVM_Task, 'duration_secs': 0.583406} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.708181] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Reconfigured VM instance instance-0000003e to attach disk [datastore1] b79f0e79-9e3a-47c7-9949-8743601ec6c5/b79f0e79-9e3a-47c7-9949-8743601ec6c5.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 801.708856] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ffc8d26-8d22-4781-a74a-e3a9a9a6de7b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.718669] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 801.718669] env[62368]: value = "task-1198448" [ 801.718669] env[62368]: _type = "Task" [ 801.718669] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.732067] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198448, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.782180] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquiring lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.784931] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.912311] env[62368]: DEBUG oslo_concurrency.lockutils [req-ebd5986a-7d48-4e0d-8f21-33c0e7fa02ec req-0a113569-931c-4e7d-bebe-5626f5b191bc service nova] Releasing lock "refresh_cache-54083854-e314-44a7-b4b0-fbef5fa6b1ef" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.925849] env[62368]: DEBUG nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.979328] env[62368]: DEBUG nova.network.neutron [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Successfully created port: b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.056500] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527cfca3-5f4d-73bb-8b20-9ed7f52d198f, 'name': SearchDatastore_Task, 'duration_secs': 0.014518} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.056839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.057256] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 54083854-e314-44a7-b4b0-fbef5fa6b1ef/54083854-e314-44a7-b4b0-fbef5fa6b1ef.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 802.057582] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb300b89-24e5-4416-b02d-23d3725edf71 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.068600] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 802.068600] env[62368]: value = "task-1198449" [ 802.068600] env[62368]: _type = "Task" [ 802.068600] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.081020] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198449, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.195053] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 802.195053] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52482d53-557e-b050-9dd6-1c8210f9799f" [ 802.195053] env[62368]: _type = "HttpNfcLease" [ 802.195053] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 802.195215] env[62368]: DEBUG oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 802.195215] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52482d53-557e-b050-9dd6-1c8210f9799f" [ 802.195215] env[62368]: _type = "HttpNfcLease" [ 802.195215] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 802.196555] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2114d9-c617-4693-922f-c73ebd0856e0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.209750] env[62368]: DEBUG oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a4e4ac-5c09-022c-3adc-c4b8e1904665/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 802.210021] env[62368]: DEBUG oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a4e4ac-5c09-022c-3adc-c4b8e1904665/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 802.288073] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198448, 'name': Rename_Task, 'duration_secs': 0.314881} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.290964] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 802.291561] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46e62788-a1ad-48a1-b778-d23b8c115ed6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.300994] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 802.300994] env[62368]: value = "task-1198450" [ 802.300994] env[62368]: _type = "Task" [ 802.300994] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.315559] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198450, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.327198] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b62e0b2c-879c-4a22-80da-3ab6d5e47e07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.458700] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449c55b0-66ad-4bf4-a90e-3b26fd6f0fd5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.476068] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db159a2a-09f3-4a15-bfb0-436c26dd9e65 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.479892] env[62368]: DEBUG nova.compute.manager [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Stashing vm_state: active {{(pid=62368) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 802.521020] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56832a42-121e-4dbc-bfdc-3a7186e646c5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.530754] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a8ec5f-615b-4638-997d-40a2521bd405 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.554280] env[62368]: DEBUG nova.compute.provider_tree [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.580752] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198449, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509006} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.581124] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 54083854-e314-44a7-b4b0-fbef5fa6b1ef/54083854-e314-44a7-b4b0-fbef5fa6b1ef.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 802.581501] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.581926] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30e4ec0a-240e-4d1d-a2df-ce62ebf87313 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.591035] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 802.591035] env[62368]: value = "task-1198451" [ 802.591035] env[62368]: _type = "Task" [ 802.591035] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.600916] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198451, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.814051] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198450, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.939205] env[62368]: DEBUG nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.967126] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.967523] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.967802] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.968122] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.968122] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.968249] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.968468] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.968698] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.968931] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.969129] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.969392] env[62368]: DEBUG nova.virt.hardware [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.970338] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430f76d2-8f69-41c5-a94d-d0cd2277cfd9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.980742] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04b1583-1cc7-4050-be14-f44dbfd17dae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.004900] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.061495] env[62368]: DEBUG nova.scheduler.client.report [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.102709] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198451, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093135} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.102994] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.103929] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef98bdb-1fcf-475b-a4bc-8fca962d37f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.129922] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 54083854-e314-44a7-b4b0-fbef5fa6b1ef/54083854-e314-44a7-b4b0-fbef5fa6b1ef.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.130429] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f6c9a70-8860-4fb4-a507-d53ba2ca6e98 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.168307] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 803.168307] env[62368]: value = "task-1198452" [ 803.168307] env[62368]: _type = "Task" [ 803.168307] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.178793] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198452, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.312943] env[62368]: DEBUG oslo_vmware.api [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198450, 'name': PowerOnVM_Task, 'duration_secs': 0.729017} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.313559] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 803.313977] env[62368]: INFO nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Took 9.58 seconds to spawn the instance on the hypervisor. [ 803.314348] env[62368]: DEBUG nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.315621] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddeafa10-ca2e-4e1c-b167-f6f44ad61a25 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.551762] env[62368]: DEBUG nova.compute.manager [req-4ecf628a-7f7a-4b5c-ad65-9718b5bdf9ad req-12b79b5c-2fa2-4657-a05e-39affaa555c4 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received event network-vif-plugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.552148] env[62368]: DEBUG oslo_concurrency.lockutils [req-4ecf628a-7f7a-4b5c-ad65-9718b5bdf9ad req-12b79b5c-2fa2-4657-a05e-39affaa555c4 service nova] Acquiring lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.552429] env[62368]: DEBUG oslo_concurrency.lockutils [req-4ecf628a-7f7a-4b5c-ad65-9718b5bdf9ad req-12b79b5c-2fa2-4657-a05e-39affaa555c4 service nova] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.552914] env[62368]: DEBUG oslo_concurrency.lockutils [req-4ecf628a-7f7a-4b5c-ad65-9718b5bdf9ad req-12b79b5c-2fa2-4657-a05e-39affaa555c4 service nova] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.553781] env[62368]: DEBUG nova.compute.manager [req-4ecf628a-7f7a-4b5c-ad65-9718b5bdf9ad req-12b79b5c-2fa2-4657-a05e-39affaa555c4 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] No waiting events found dispatching network-vif-plugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.554141] env[62368]: WARNING nova.compute.manager [req-4ecf628a-7f7a-4b5c-ad65-9718b5bdf9ad req-12b79b5c-2fa2-4657-a05e-39affaa555c4 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received unexpected event network-vif-plugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e for instance with vm_state building and task_state spawning. [ 803.568097] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.664s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.568845] env[62368]: DEBUG nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.571858] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.260s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.573847] env[62368]: INFO nova.compute.claims [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.674583] env[62368]: DEBUG nova.network.neutron [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Successfully updated port: b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.683335] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.836023] env[62368]: INFO nova.compute.manager [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Took 40.82 seconds to build instance. [ 804.083903] env[62368]: DEBUG nova.compute.utils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.089733] env[62368]: DEBUG nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 804.090094] env[62368]: DEBUG nova.network.neutron [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 804.136906] env[62368]: DEBUG nova.policy [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3fbd607d8b74e22987a89cd0da5f065', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e31b8f1352574bb7808b06e732da7e4e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 804.181153] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.181153] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.184198] env[62368]: DEBUG nova.network.neutron [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.192060] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198452, 'name': ReconfigVM_Task, 'duration_secs': 0.7033} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.192666] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 54083854-e314-44a7-b4b0-fbef5fa6b1ef/54083854-e314-44a7-b4b0-fbef5fa6b1ef.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.197098] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1e3434c-de2e-4132-a491-f7a84e1b3b71 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.212887] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 804.212887] env[62368]: value = "task-1198453" [ 804.212887] env[62368]: _type = "Task" [ 804.212887] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.223703] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198453, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.337737] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5e6b4d76-6266-4242-97a0-1b1ec144ee9f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.419s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.541138] env[62368]: DEBUG nova.network.neutron [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Successfully created port: 46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.589898] env[62368]: DEBUG nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.725031] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198453, 'name': Rename_Task, 'duration_secs': 0.166392} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.725031] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 804.727033] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c0c5236-a387-4f63-b129-398c630b6fc4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.737514] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 804.737514] env[62368]: value = "task-1198454" [ 804.737514] env[62368]: _type = "Task" [ 804.737514] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.745516] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198454, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.746439] env[62368]: DEBUG nova.network.neutron [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.843839] env[62368]: DEBUG nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 804.977527] env[62368]: DEBUG nova.network.neutron [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating instance_info_cache with network_info: [{"id": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "address": "fa:16:3e:d1:c4:2e", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ebe893-6b", "ovs_interfaceid": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.017331] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd255d07-dc67-4002-8cbf-06d5172d2338 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.028197] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df4784f-2260-4786-9d29-431b4ce16130 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.060742] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f96db0c-e43e-4f73-b66b-5a79bdf87ad1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.069834] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2c4442-3704-46aa-8f28-d0c9a4b7ac3a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.087019] env[62368]: DEBUG nova.compute.provider_tree [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.246903] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198454, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.373911] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.456796] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "5ed9cee5-8cc8-4b70-96a9-737019363638" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.457160] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "5ed9cee5-8cc8-4b70-96a9-737019363638" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.480599] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.480930] env[62368]: DEBUG nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Instance network_info: |[{"id": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "address": "fa:16:3e:d1:c4:2e", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ebe893-6b", "ovs_interfaceid": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.481630] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:c4:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7150f662-0cf1-44f9-ae14-d70f479649b6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2ebe893-6ba3-40da-8def-fb6cf1ab442e', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.490025] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating folder: Project (e250606f3df240eb93ce45925b00ad43). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 805.490277] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-154848c4-71a4-4bdc-b73c-c7c01ee5dc3b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.502884] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Created folder: Project (e250606f3df240eb93ce45925b00ad43) in parent group-v259706. [ 805.503306] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating folder: Instances. Parent ref: group-v259797. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 805.503392] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd933c00-9c03-4890-80cd-c564597eff1d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.513999] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Created folder: Instances in parent group-v259797. [ 805.514291] env[62368]: DEBUG oslo.service.loopingcall [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.514502] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 805.514722] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c5777f2-28cf-4afe-a6cf-c081cc473878 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.533843] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.533843] env[62368]: value = "task-1198457" [ 805.533843] env[62368]: _type = "Task" [ 805.533843] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.543480] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198457, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.580641] env[62368]: DEBUG nova.compute.manager [req-03466740-fd87-4385-a38b-1265e2ec7f04 req-453e8d3a-5297-4ad4-b535-fb2237d452c3 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received event network-changed-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.580860] env[62368]: DEBUG nova.compute.manager [req-03466740-fd87-4385-a38b-1265e2ec7f04 req-453e8d3a-5297-4ad4-b535-fb2237d452c3 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Refreshing instance network info cache due to event network-changed-b2ebe893-6ba3-40da-8def-fb6cf1ab442e. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.581072] env[62368]: DEBUG oslo_concurrency.lockutils [req-03466740-fd87-4385-a38b-1265e2ec7f04 req-453e8d3a-5297-4ad4-b535-fb2237d452c3 service nova] Acquiring lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.581251] env[62368]: DEBUG oslo_concurrency.lockutils [req-03466740-fd87-4385-a38b-1265e2ec7f04 req-453e8d3a-5297-4ad4-b535-fb2237d452c3 service nova] Acquired lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.581411] env[62368]: DEBUG nova.network.neutron [req-03466740-fd87-4385-a38b-1265e2ec7f04 req-453e8d3a-5297-4ad4-b535-fb2237d452c3 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Refreshing network info cache for port b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.589775] env[62368]: DEBUG nova.scheduler.client.report [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.607426] env[62368]: DEBUG nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.634996] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.635311] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.635489] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.635677] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.635826] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.636065] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.636351] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.636544] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.636734] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.636904] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.637109] env[62368]: DEBUG nova.virt.hardware [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.638046] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f91401-03f8-4008-b276-80cacd74d42e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.647614] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036c936b-7071-4a21-8826-d440ff280842 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.747477] env[62368]: DEBUG oslo_vmware.api [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198454, 'name': PowerOnVM_Task, 'duration_secs': 0.588526} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.747773] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 805.747987] env[62368]: INFO nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Took 8.40 seconds to spawn the instance on the hypervisor. [ 805.748202] env[62368]: DEBUG nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.748994] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead7d8d9-1aaf-4f4f-97e0-3f1796d7dc52 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.047309] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198457, 'name': CreateVM_Task, 'duration_secs': 0.486572} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.047551] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 806.049067] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.050034] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.050034] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.050236] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dc2a613-0021-4ad0-a356-c9a7b3494c29 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.058579] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 806.058579] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5238fef9-8a1f-b52b-0a6b-04f1364179fc" [ 806.058579] env[62368]: _type = "Task" [ 806.058579] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.069640] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5238fef9-8a1f-b52b-0a6b-04f1364179fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.094683] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.095321] env[62368]: DEBUG nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 806.101562] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.636s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.101562] env[62368]: INFO nova.compute.claims [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.268286] env[62368]: INFO nova.compute.manager [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Took 36.52 seconds to build instance. [ 806.303522] env[62368]: DEBUG nova.network.neutron [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Successfully updated port: 46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.385024] env[62368]: DEBUG nova.network.neutron [req-03466740-fd87-4385-a38b-1265e2ec7f04 req-453e8d3a-5297-4ad4-b535-fb2237d452c3 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updated VIF entry in instance network info cache for port b2ebe893-6ba3-40da-8def-fb6cf1ab442e. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 806.385442] env[62368]: DEBUG nova.network.neutron [req-03466740-fd87-4385-a38b-1265e2ec7f04 req-453e8d3a-5297-4ad4-b535-fb2237d452c3 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating instance_info_cache with network_info: [{"id": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "address": "fa:16:3e:d1:c4:2e", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ebe893-6b", "ovs_interfaceid": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.573603] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5238fef9-8a1f-b52b-0a6b-04f1364179fc, 'name': SearchDatastore_Task, 'duration_secs': 0.012982} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.573745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.574037] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.574382] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.574588] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.574838] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.575223] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ac856c8-a22e-42d1-9c4d-be72b06f5a58 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.585646] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.585844] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.586621] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e706f8a9-473b-4dde-9874-66f412a79bde {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.593099] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 806.593099] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52d88141-1417-e294-aff9-63e5695bae7a" [ 806.593099] env[62368]: _type = "Task" [ 806.593099] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.602993] env[62368]: DEBUG nova.compute.utils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.604365] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d88141-1417-e294-aff9-63e5695bae7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.604659] env[62368]: DEBUG nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 806.604840] env[62368]: DEBUG nova.network.neutron [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 806.648968] env[62368]: DEBUG nova.policy [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f3afe492e7e47f28c3c8dad61f819d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81b69987192149ddb6dfcafc1f7770fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.775613] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc17b9aa-fe08-4483-b3a6-bb3c43ce497d tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.118s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.806513] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.806513] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.806755] env[62368]: DEBUG nova.network.neutron [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.892865] env[62368]: DEBUG oslo_concurrency.lockutils [req-03466740-fd87-4385-a38b-1265e2ec7f04 req-453e8d3a-5297-4ad4-b535-fb2237d452c3 service nova] Releasing lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.914649] env[62368]: DEBUG nova.network.neutron [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Successfully created port: 3ba9f19e-f789-4e0a-b733-bba38fc85baf {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.108116] env[62368]: DEBUG nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 807.111837] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d88141-1417-e294-aff9-63e5695bae7a, 'name': SearchDatastore_Task, 'duration_secs': 0.009909} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.116367] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3f95559-bb86-46c1-8998-0e49b3d9a8ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.125339] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 807.125339] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52d8b018-e393-45d8-281c-6334eaac60d3" [ 807.125339] env[62368]: _type = "Task" [ 807.125339] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.140314] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d8b018-e393-45d8-281c-6334eaac60d3, 'name': SearchDatastore_Task, 'duration_secs': 0.010508} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.140661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.141094] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6/eea21546-fbbf-4440-829c-8583c4ccabb6.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 807.141536] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6dc12564-7459-4b20-8631-f9c121910e75 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.166281] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 807.166281] env[62368]: value = "task-1198458" [ 807.166281] env[62368]: _type = "Task" [ 807.166281] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.176291] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.278621] env[62368]: DEBUG nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.353446] env[62368]: DEBUG nova.network.neutron [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.422181] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.422560] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.422902] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.423151] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.423414] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.429799] env[62368]: INFO nova.compute.manager [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Terminating instance [ 807.433164] env[62368]: DEBUG nova.compute.manager [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.433402] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 807.434395] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735f63a9-b664-4cdc-8a7d-6428cf6c3374 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.448386] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.448847] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da6d8ef7-11a0-480e-891f-69eeb4a2a0de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.459458] env[62368]: DEBUG oslo_vmware.api [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 807.459458] env[62368]: value = "task-1198459" [ 807.459458] env[62368]: _type = "Task" [ 807.459458] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.469919] env[62368]: DEBUG oslo_vmware.api [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198459, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.578107] env[62368]: DEBUG nova.network.neutron [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updating instance_info_cache with network_info: [{"id": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "address": "fa:16:3e:23:0d:cc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46fba643-18", "ovs_interfaceid": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.582935] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2068201-8227-484a-8901-9d65c9f137b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.593829] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76293e89-ba35-44d4-9844-4edf5c48de55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.640263] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884e33f7-c396-4341-bf52-287335195a1c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.649133] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0922ad1e-58fc-4820-8611-aebcc3913c5c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.665697] env[62368]: DEBUG nova.compute.provider_tree [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.677103] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510287} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.678475] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6/eea21546-fbbf-4440-829c-8583c4ccabb6.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.678708] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.678996] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09a2235f-124f-4d14-857f-cb2c68f37d50 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.688247] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 807.688247] env[62368]: value = "task-1198460" [ 807.688247] env[62368]: _type = "Task" [ 807.688247] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.699551] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198460, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.780732] env[62368]: DEBUG nova.compute.manager [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Received event network-vif-plugged-46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.781176] env[62368]: DEBUG oslo_concurrency.lockutils [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.781527] env[62368]: DEBUG oslo_concurrency.lockutils [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.781757] env[62368]: DEBUG oslo_concurrency.lockutils [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.781984] env[62368]: DEBUG nova.compute.manager [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] No waiting events found dispatching network-vif-plugged-46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 807.782262] env[62368]: WARNING nova.compute.manager [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Received unexpected event network-vif-plugged-46fba643-185a-4fb2-ab24-dd4350aa5ac4 for instance with vm_state building and task_state spawning. [ 807.782473] env[62368]: DEBUG nova.compute.manager [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Received event network-changed-46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.782668] env[62368]: DEBUG nova.compute.manager [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Refreshing instance network info cache due to event network-changed-46fba643-185a-4fb2-ab24-dd4350aa5ac4. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 807.782947] env[62368]: DEBUG oslo_concurrency.lockutils [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] Acquiring lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.799307] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.970325] env[62368]: DEBUG oslo_vmware.api [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198459, 'name': PowerOffVM_Task, 'duration_secs': 0.262385} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.970596] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 807.970770] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 807.971041] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5218511c-b523-4532-bcd0-00326f893048 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.087861] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Releasing lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.088189] env[62368]: DEBUG nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Instance network_info: |[{"id": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "address": "fa:16:3e:23:0d:cc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46fba643-18", "ovs_interfaceid": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.088656] env[62368]: DEBUG oslo_concurrency.lockutils [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] Acquired lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.088819] env[62368]: DEBUG nova.network.neutron [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Refreshing network info cache for port 46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.092071] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:0d:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '975b168a-03e5-449d-95ac-4d51ba027242', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46fba643-185a-4fb2-ab24-dd4350aa5ac4', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.099551] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Creating folder: Project (e31b8f1352574bb7808b06e732da7e4e). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 808.102199] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af965d8d-b856-4e58-9284-95330c3ad501 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.104407] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 808.104684] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 808.104943] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Deleting the datastore file [datastore1] 54083854-e314-44a7-b4b0-fbef5fa6b1ef {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.105559] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eda197f9-ee53-4115-8fea-b00a9ab7890e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.114303] env[62368]: DEBUG oslo_vmware.api [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for the task: (returnval){ [ 808.114303] env[62368]: value = "task-1198463" [ 808.114303] env[62368]: _type = "Task" [ 808.114303] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.119200] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Created folder: Project (e31b8f1352574bb7808b06e732da7e4e) in parent group-v259706. [ 808.119474] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Creating folder: Instances. Parent ref: group-v259800. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 808.120182] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc1880f9-8f68-47b7-bd45-1d7d3e8c9658 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.127964] env[62368]: DEBUG oslo_vmware.api [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198463, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.130296] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Created folder: Instances in parent group-v259800. [ 808.130535] env[62368]: DEBUG oslo.service.loopingcall [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.130738] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 808.130954] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d67ace9c-6b1f-4dcd-a605-90a88d499143 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.146569] env[62368]: DEBUG nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 808.169819] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.169819] env[62368]: value = "task-1198465" [ 808.169819] env[62368]: _type = "Task" [ 808.169819] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.172263] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='21ccd447e67e89acac22da9e73694698',container_format='bare',created_at=2024-10-08T23:17:02Z,direct_url=,disk_format='vmdk',id=6281ea63-c477-4e65-8fac-7b43765b0423,min_disk=1,min_ram=0,name='tempest-test-snap-1499571954',owner='81b69987192149ddb6dfcafc1f7770fb',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-08T23:17:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 808.172520] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 808.172662] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 808.172858] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 808.173016] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 808.173182] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 808.173396] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 808.173555] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 808.173722] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 808.173940] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 808.174145] env[62368]: DEBUG nova.virt.hardware [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.175080] env[62368]: DEBUG nova.scheduler.client.report [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.179197] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b01be7d-0bf1-43e0-91a7-641b4fe2816c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.197465] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ab6732-910b-4134-b890-a07ed6ae0396 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.201935] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198465, 'name': CreateVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.210009] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198460, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09626} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.219187] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.220711] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567c73a3-a39a-4cb6-a5e3-22f2cbbad8ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.248994] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6/eea21546-fbbf-4440-829c-8583c4ccabb6.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.248994] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e66061fa-0a2d-47ef-bfaf-cd84df01a66f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.271787] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 808.271787] env[62368]: value = "task-1198466" [ 808.271787] env[62368]: _type = "Task" [ 808.271787] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.283381] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198466, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.615839] env[62368]: DEBUG nova.compute.manager [req-62716df9-163d-46e6-855d-4310032eb6b9 req-28856a06-630c-47e8-8ff9-485aef5a518f service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Received event network-vif-plugged-3ba9f19e-f789-4e0a-b733-bba38fc85baf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.615839] env[62368]: DEBUG oslo_concurrency.lockutils [req-62716df9-163d-46e6-855d-4310032eb6b9 req-28856a06-630c-47e8-8ff9-485aef5a518f service nova] Acquiring lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.615839] env[62368]: DEBUG oslo_concurrency.lockutils [req-62716df9-163d-46e6-855d-4310032eb6b9 req-28856a06-630c-47e8-8ff9-485aef5a518f service nova] Lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.615839] env[62368]: DEBUG oslo_concurrency.lockutils [req-62716df9-163d-46e6-855d-4310032eb6b9 req-28856a06-630c-47e8-8ff9-485aef5a518f service nova] Lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.616042] env[62368]: DEBUG nova.compute.manager [req-62716df9-163d-46e6-855d-4310032eb6b9 req-28856a06-630c-47e8-8ff9-485aef5a518f service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] No waiting events found dispatching network-vif-plugged-3ba9f19e-f789-4e0a-b733-bba38fc85baf {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.616158] env[62368]: WARNING nova.compute.manager [req-62716df9-163d-46e6-855d-4310032eb6b9 req-28856a06-630c-47e8-8ff9-485aef5a518f service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Received unexpected event network-vif-plugged-3ba9f19e-f789-4e0a-b733-bba38fc85baf for instance with vm_state building and task_state spawning. [ 808.635617] env[62368]: DEBUG oslo_vmware.api [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Task: {'id': task-1198463, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171475} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.635617] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.635617] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 808.635617] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 808.635617] env[62368]: INFO nova.compute.manager [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Took 1.20 seconds to destroy the instance on the hypervisor. [ 808.636886] env[62368]: DEBUG oslo.service.loopingcall [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.640623] env[62368]: DEBUG nova.compute.manager [-] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.640623] env[62368]: DEBUG nova.network.neutron [-] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.682918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.683929] env[62368]: DEBUG nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 808.696606] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.400s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.697209] env[62368]: DEBUG nova.objects.instance [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lazy-loading 'resources' on Instance uuid 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.698590] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198465, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.727203] env[62368]: DEBUG nova.network.neutron [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Successfully updated port: 3ba9f19e-f789-4e0a-b733-bba38fc85baf {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 808.783613] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.187499] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198465, 'name': CreateVM_Task, 'duration_secs': 0.634564} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.187848] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 809.188504] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.188601] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.188913] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.189217] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de1d4cc3-7bfa-4dc8-bb4d-d86e3092f002 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.194922] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 809.194922] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522ce788-0ec2-fb09-7b4c-3abcacc184ff" [ 809.194922] env[62368]: _type = "Task" [ 809.194922] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.206238] env[62368]: DEBUG nova.compute.utils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.207431] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522ce788-0ec2-fb09-7b4c-3abcacc184ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.211604] env[62368]: DEBUG nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.211780] env[62368]: DEBUG nova.network.neutron [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.232642] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "refresh_cache-0b861a8b-dcad-46ad-90ef-a46450c5cd51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.232799] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "refresh_cache-0b861a8b-dcad-46ad-90ef-a46450c5cd51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.232954] env[62368]: DEBUG nova.network.neutron [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.273876] env[62368]: DEBUG nova.policy [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6972fdf35fc4fe9afa65cdfd0d46568', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb560320635f493fbe4308af1623e8ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 809.276769] env[62368]: DEBUG nova.network.neutron [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updated VIF entry in instance network info cache for port 46fba643-185a-4fb2-ab24-dd4350aa5ac4. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.278423] env[62368]: DEBUG nova.network.neutron [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updating instance_info_cache with network_info: [{"id": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "address": "fa:16:3e:23:0d:cc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46fba643-18", "ovs_interfaceid": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.305194] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198466, 'name': ReconfigVM_Task, 'duration_secs': 0.59638} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.312511] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Reconfigured VM instance instance-00000040 to attach disk [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6/eea21546-fbbf-4440-829c-8583c4ccabb6.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.312511] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6bd3a695-5867-4199-9853-622a3352487b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.324025] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 809.324025] env[62368]: value = "task-1198467" [ 809.324025] env[62368]: _type = "Task" [ 809.324025] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.340472] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198467, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.669942] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb86903-ccc1-4231-b2e8-f42c921294d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.678722] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ec7e8b-f995-4291-b491-8560670d74b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.713285] env[62368]: DEBUG nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 809.716859] env[62368]: DEBUG nova.network.neutron [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Successfully created port: 0f20ad94-5a78-40ab-9ce7-1a56c05d20bc {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 809.722461] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1b9be5-5158-4bbd-a83d-6a4b1bbcd477 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.737347] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522ce788-0ec2-fb09-7b4c-3abcacc184ff, 'name': SearchDatastore_Task, 'duration_secs': 0.018131} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.738646] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e8fea0-5661-4f4a-a8f6-57fb65545788 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.742719] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.743070] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.743522] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.743595] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.743826] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.744557] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b093313d-3871-40a1-a556-47f326120ad0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.758111] env[62368]: DEBUG nova.compute.provider_tree [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.760474] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.760688] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 809.762697] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca297373-2651-4c77-80aa-7c0acc88b51a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.769877] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 809.769877] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52b960fa-10c5-6015-aa73-6e6169a873bf" [ 809.769877] env[62368]: _type = "Task" [ 809.769877] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.779535] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b960fa-10c5-6015-aa73-6e6169a873bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.780361] env[62368]: DEBUG nova.network.neutron [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 809.792516] env[62368]: DEBUG oslo_concurrency.lockutils [req-7931dd24-97e2-4ed3-b463-024d2c4983f2 req-f462062b-e16f-4dcb-9134-40b3b6cefdbe service nova] Releasing lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.818028] env[62368]: DEBUG nova.compute.manager [req-64742e0a-f3a6-4f4b-89a9-b7708ca71bda req-f50a1577-36ad-4798-bde2-8c215bb88ccd service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Received event network-vif-deleted-31693720-039f-4059-b582-9c57cbe54814 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.818136] env[62368]: INFO nova.compute.manager [req-64742e0a-f3a6-4f4b-89a9-b7708ca71bda req-f50a1577-36ad-4798-bde2-8c215bb88ccd service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Neutron deleted interface 31693720-039f-4059-b582-9c57cbe54814; detaching it from the instance and deleting it from the info cache [ 809.818306] env[62368]: DEBUG nova.network.neutron [req-64742e0a-f3a6-4f4b-89a9-b7708ca71bda req-f50a1577-36ad-4798-bde2-8c215bb88ccd service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.836443] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198467, 'name': Rename_Task, 'duration_secs': 0.319092} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.837250] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 809.837509] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e36fea91-af39-4335-8a35-b1a267bdcac6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.844998] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 809.844998] env[62368]: value = "task-1198468" [ 809.844998] env[62368]: _type = "Task" [ 809.844998] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.855123] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198468, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.899405] env[62368]: DEBUG nova.network.neutron [-] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.940368] env[62368]: DEBUG nova.network.neutron [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Updating instance_info_cache with network_info: [{"id": "3ba9f19e-f789-4e0a-b733-bba38fc85baf", "address": "fa:16:3e:98:66:90", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ba9f19e-f7", "ovs_interfaceid": "3ba9f19e-f789-4e0a-b733-bba38fc85baf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.263381] env[62368]: DEBUG nova.scheduler.client.report [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.281903] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b960fa-10c5-6015-aa73-6e6169a873bf, 'name': SearchDatastore_Task, 'duration_secs': 0.010606} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.282804] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-299d50f3-fd3c-47fd-9671-a7836622ce3d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.290363] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 810.290363] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52145357-f365-87c4-a86e-62a570ffe828" [ 810.290363] env[62368]: _type = "Task" [ 810.290363] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.299240] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52145357-f365-87c4-a86e-62a570ffe828, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.323595] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00ed86e7-28d8-481a-85f9-d1fa5ba4f8bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.335745] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6211ea95-b441-45af-9fa3-062b794d2338 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.359152] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198468, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.371186] env[62368]: DEBUG nova.compute.manager [req-64742e0a-f3a6-4f4b-89a9-b7708ca71bda req-f50a1577-36ad-4798-bde2-8c215bb88ccd service nova] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Detach interface failed, port_id=31693720-039f-4059-b582-9c57cbe54814, reason: Instance 54083854-e314-44a7-b4b0-fbef5fa6b1ef could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 810.402052] env[62368]: INFO nova.compute.manager [-] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Took 1.76 seconds to deallocate network for instance. [ 810.443304] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "refresh_cache-0b861a8b-dcad-46ad-90ef-a46450c5cd51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.443670] env[62368]: DEBUG nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Instance network_info: |[{"id": "3ba9f19e-f789-4e0a-b733-bba38fc85baf", "address": "fa:16:3e:98:66:90", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ba9f19e-f7", "ovs_interfaceid": "3ba9f19e-f789-4e0a-b733-bba38fc85baf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 810.444251] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:66:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ba9f19e-f789-4e0a-b733-bba38fc85baf', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.452115] env[62368]: DEBUG oslo.service.loopingcall [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.452778] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 810.452926] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ff5c651-c727-4507-b3e0-63653c3f6e55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.475647] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.475647] env[62368]: value = "task-1198469" [ 810.475647] env[62368]: _type = "Task" [ 810.475647] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.484578] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198469, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.641462] env[62368]: DEBUG nova.compute.manager [req-67b2175f-48e7-4f6a-8060-bfae33950816 req-5d57ffb6-d636-492d-9c28-a295e33785fd service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Received event network-changed-3ba9f19e-f789-4e0a-b733-bba38fc85baf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.641686] env[62368]: DEBUG nova.compute.manager [req-67b2175f-48e7-4f6a-8060-bfae33950816 req-5d57ffb6-d636-492d-9c28-a295e33785fd service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Refreshing instance network info cache due to event network-changed-3ba9f19e-f789-4e0a-b733-bba38fc85baf. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 810.642111] env[62368]: DEBUG oslo_concurrency.lockutils [req-67b2175f-48e7-4f6a-8060-bfae33950816 req-5d57ffb6-d636-492d-9c28-a295e33785fd service nova] Acquiring lock "refresh_cache-0b861a8b-dcad-46ad-90ef-a46450c5cd51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.642310] env[62368]: DEBUG oslo_concurrency.lockutils [req-67b2175f-48e7-4f6a-8060-bfae33950816 req-5d57ffb6-d636-492d-9c28-a295e33785fd service nova] Acquired lock "refresh_cache-0b861a8b-dcad-46ad-90ef-a46450c5cd51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.642493] env[62368]: DEBUG nova.network.neutron [req-67b2175f-48e7-4f6a-8060-bfae33950816 req-5d57ffb6-d636-492d-9c28-a295e33785fd service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Refreshing network info cache for port 3ba9f19e-f789-4e0a-b733-bba38fc85baf {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 810.733140] env[62368]: DEBUG nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 810.756271] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.757022] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.757022] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.757022] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.757180] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.759940] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.759940] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.759940] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.759940] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.759940] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.759940] env[62368]: DEBUG nova.virt.hardware [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.759940] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df604b2-9b76-458a-bf72-eb1fbe589d9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.772291] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082b9522-2678-4bfe-b2e2-0432505f5baf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.774247] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.078s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.777993] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.953s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.778351] env[62368]: DEBUG nova.objects.instance [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lazy-loading 'resources' on Instance uuid 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 810.805444] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52145357-f365-87c4-a86e-62a570ffe828, 'name': SearchDatastore_Task, 'duration_secs': 0.011935} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.805444] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.805558] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 58157ab1-80a4-427c-812b-f6fde1f8db68/58157ab1-80a4-427c-812b-f6fde1f8db68.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 810.806651] env[62368]: INFO nova.scheduler.client.report [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Deleted allocations for instance 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072 [ 810.807775] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12129eb0-871e-4be4-b14e-31ece298f589 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.820441] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 810.820441] env[62368]: value = "task-1198470" [ 810.820441] env[62368]: _type = "Task" [ 810.820441] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.830672] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.862071] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198468, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.908744] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.992009] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198469, 'name': CreateVM_Task, 'duration_secs': 0.351648} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.992009] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 810.992638] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.992808] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.993230] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.993508] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9bde6d7-039a-4a63-a650-263e6400ada1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.999958] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 810.999958] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527da3a5-59eb-b51c-eb28-994282851da9" [ 810.999958] env[62368]: _type = "Task" [ 810.999958] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.013156] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527da3a5-59eb-b51c-eb28-994282851da9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.318407] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3d80bdf8-933a-4646-b2ae-94e824726518 tempest-AttachInterfacesUnderV243Test-937372910 tempest-AttachInterfacesUnderV243Test-937372910-project-member] Lock "8b7ad7ed-8f2d-4c76-b335-4a2ce5327072" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.024s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.332194] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198470, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498279} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.334911] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 58157ab1-80a4-427c-812b-f6fde1f8db68/58157ab1-80a4-427c-812b-f6fde1f8db68.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 811.335100] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.336077] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64748b5f-382b-45fd-b05a-19ac8aeb1eb6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.344840] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 811.344840] env[62368]: value = "task-1198471" [ 811.344840] env[62368]: _type = "Task" [ 811.344840] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.361043] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198471, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.369044] env[62368]: DEBUG oslo_vmware.api [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198468, 'name': PowerOnVM_Task, 'duration_secs': 1.436356} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.369044] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 811.369044] env[62368]: INFO nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Took 8.43 seconds to spawn the instance on the hypervisor. [ 811.369044] env[62368]: DEBUG nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.369637] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a658ae1d-f5ff-47e7-9f0d-4d4c9f9a74b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.417202] env[62368]: DEBUG nova.network.neutron [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Successfully updated port: 0f20ad94-5a78-40ab-9ce7-1a56c05d20bc {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.497797] env[62368]: DEBUG nova.network.neutron [req-67b2175f-48e7-4f6a-8060-bfae33950816 req-5d57ffb6-d636-492d-9c28-a295e33785fd service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Updated VIF entry in instance network info cache for port 3ba9f19e-f789-4e0a-b733-bba38fc85baf. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 811.498237] env[62368]: DEBUG nova.network.neutron [req-67b2175f-48e7-4f6a-8060-bfae33950816 req-5d57ffb6-d636-492d-9c28-a295e33785fd service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Updating instance_info_cache with network_info: [{"id": "3ba9f19e-f789-4e0a-b733-bba38fc85baf", "address": "fa:16:3e:98:66:90", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ba9f19e-f7", "ovs_interfaceid": "3ba9f19e-f789-4e0a-b733-bba38fc85baf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.515185] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.515560] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Processing image 6281ea63-c477-4e65-8fac-7b43765b0423 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.515866] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423/6281ea63-c477-4e65-8fac-7b43765b0423.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.516092] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423/6281ea63-c477-4e65-8fac-7b43765b0423.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.516426] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.517140] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb2358b0-5a41-43cf-9a01-735d94cb10b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.527593] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.527778] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 811.530871] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49cf9726-8310-4a95-b6f5-23a71dc6051c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.537308] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 811.537308] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521c2736-3477-9f29-83de-af9cb2b547e6" [ 811.537308] env[62368]: _type = "Task" [ 811.537308] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.545913] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521c2736-3477-9f29-83de-af9cb2b547e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.663742] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eeb4ad9-ef06-4620-a925-d4cae1830e4b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.671501] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd392a34-b860-427a-9b62-d76ec8b986f4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.702930] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e6e238-9dfb-4bed-a1c6-4c8fbe1609b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.710810] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b576dc8-3b0c-4053-8ad2-369c35d86dfd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.725637] env[62368]: DEBUG nova.compute.provider_tree [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.855677] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198471, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069718} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.855969] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.856780] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f50df82-c814-47ad-ab78-67543aec67ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.879712] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 58157ab1-80a4-427c-812b-f6fde1f8db68/58157ab1-80a4-427c-812b-f6fde1f8db68.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.880779] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ac08221-26cd-42c3-b634-8cf91c56fc77 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.902628] env[62368]: INFO nova.compute.manager [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Took 35.21 seconds to build instance. [ 811.910301] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 811.910301] env[62368]: value = "task-1198472" [ 811.910301] env[62368]: _type = "Task" [ 811.910301] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.920545] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "refresh_cache-d865d433-2341-4f6c-b840-609530ae2e51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.920689] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "refresh_cache-d865d433-2341-4f6c-b840-609530ae2e51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.920847] env[62368]: DEBUG nova.network.neutron [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.922261] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198472, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.003368] env[62368]: DEBUG oslo_concurrency.lockutils [req-67b2175f-48e7-4f6a-8060-bfae33950816 req-5d57ffb6-d636-492d-9c28-a295e33785fd service nova] Releasing lock "refresh_cache-0b861a8b-dcad-46ad-90ef-a46450c5cd51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.049338] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Preparing fetch location {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 812.049606] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Fetch image to [datastore1] OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0/OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0.vmdk {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 812.049792] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Downloading stream optimized image 6281ea63-c477-4e65-8fac-7b43765b0423 to [datastore1] OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0/OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0.vmdk on the data store datastore1 as vApp {{(pid=62368) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 812.050073] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Downloading image file data 6281ea63-c477-4e65-8fac-7b43765b0423 to the ESX as VM named 'OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0' {{(pid=62368) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 812.133907] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 812.133907] env[62368]: value = "resgroup-9" [ 812.133907] env[62368]: _type = "ResourcePool" [ 812.133907] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 812.134244] env[62368]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-4d317dea-8bbd-48f9-9855-80c3b0631f66 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.167381] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lease: (returnval){ [ 812.167381] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521c8998-6b79-8946-fdf3-7a9e7f7e6444" [ 812.167381] env[62368]: _type = "HttpNfcLease" [ 812.167381] env[62368]: } obtained for vApp import into resource pool (val){ [ 812.167381] env[62368]: value = "resgroup-9" [ 812.167381] env[62368]: _type = "ResourcePool" [ 812.167381] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 812.167828] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the lease: (returnval){ [ 812.167828] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521c8998-6b79-8946-fdf3-7a9e7f7e6444" [ 812.167828] env[62368]: _type = "HttpNfcLease" [ 812.167828] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 812.176220] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 812.176220] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521c8998-6b79-8946-fdf3-7a9e7f7e6444" [ 812.176220] env[62368]: _type = "HttpNfcLease" [ 812.176220] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 812.229523] env[62368]: DEBUG nova.scheduler.client.report [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.404745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4243a57-ac01-4ffd-adcb-70d7336d8046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.898s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.422249] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198472, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.479655] env[62368]: DEBUG nova.network.neutron [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.677836] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 812.677836] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521c8998-6b79-8946-fdf3-7a9e7f7e6444" [ 812.677836] env[62368]: _type = "HttpNfcLease" [ 812.677836] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 812.693588] env[62368]: DEBUG nova.network.neutron [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Updating instance_info_cache with network_info: [{"id": "0f20ad94-5a78-40ab-9ce7-1a56c05d20bc", "address": "fa:16:3e:08:b6:3a", "network": {"id": "e89b0140-71be-4766-9ec4-c37da482a0e3", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2110723737-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb560320635f493fbe4308af1623e8ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f20ad94-5a", "ovs_interfaceid": "0f20ad94-5a78-40ab-9ce7-1a56c05d20bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.735193] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.957s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.738095] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.615s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.738378] env[62368]: DEBUG nova.objects.instance [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lazy-loading 'resources' on Instance uuid 300c6c32-b365-40c2-bc7d-6ce028f10476 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.758316] env[62368]: INFO nova.scheduler.client.report [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Deleted allocations for instance 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3 [ 812.854876] env[62368]: DEBUG nova.compute.manager [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Received event network-vif-plugged-0f20ad94-5a78-40ab-9ce7-1a56c05d20bc {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.855057] env[62368]: DEBUG oslo_concurrency.lockutils [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] Acquiring lock "d865d433-2341-4f6c-b840-609530ae2e51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.855391] env[62368]: DEBUG oslo_concurrency.lockutils [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] Lock "d865d433-2341-4f6c-b840-609530ae2e51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.855515] env[62368]: DEBUG oslo_concurrency.lockutils [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] Lock "d865d433-2341-4f6c-b840-609530ae2e51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.855740] env[62368]: DEBUG nova.compute.manager [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] No waiting events found dispatching network-vif-plugged-0f20ad94-5a78-40ab-9ce7-1a56c05d20bc {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.855953] env[62368]: WARNING nova.compute.manager [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Received unexpected event network-vif-plugged-0f20ad94-5a78-40ab-9ce7-1a56c05d20bc for instance with vm_state building and task_state spawning. [ 812.856034] env[62368]: DEBUG nova.compute.manager [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Received event network-changed-0f20ad94-5a78-40ab-9ce7-1a56c05d20bc {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.856404] env[62368]: DEBUG nova.compute.manager [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Refreshing instance network info cache due to event network-changed-0f20ad94-5a78-40ab-9ce7-1a56c05d20bc. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.856600] env[62368]: DEBUG oslo_concurrency.lockutils [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] Acquiring lock "refresh_cache-d865d433-2341-4f6c-b840-609530ae2e51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.909332] env[62368]: DEBUG nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.929600] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198472, 'name': ReconfigVM_Task, 'duration_secs': 0.573968} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.929907] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 58157ab1-80a4-427c-812b-f6fde1f8db68/58157ab1-80a4-427c-812b-f6fde1f8db68.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.930817] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a315152b-6e3c-4a86-b627-268f4e3f64cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.939451] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 812.939451] env[62368]: value = "task-1198474" [ 812.939451] env[62368]: _type = "Task" [ 812.939451] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.949742] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198474, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.177668] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 813.177668] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521c8998-6b79-8946-fdf3-7a9e7f7e6444" [ 813.177668] env[62368]: _type = "HttpNfcLease" [ 813.177668] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 813.177974] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 813.177974] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521c8998-6b79-8946-fdf3-7a9e7f7e6444" [ 813.177974] env[62368]: _type = "HttpNfcLease" [ 813.177974] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 813.178729] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd46b020-6513-4bc0-93bd-88236039f22f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.187395] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520adb14-6d7a-5cbd-4ea1-f1b8c4b8280b/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 813.187588] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520adb14-6d7a-5cbd-4ea1-f1b8c4b8280b/disk-0.vmdk. {{(pid=62368) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 813.247909] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "refresh_cache-d865d433-2341-4f6c-b840-609530ae2e51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.248508] env[62368]: DEBUG nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Instance network_info: |[{"id": "0f20ad94-5a78-40ab-9ce7-1a56c05d20bc", "address": "fa:16:3e:08:b6:3a", "network": {"id": "e89b0140-71be-4766-9ec4-c37da482a0e3", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2110723737-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb560320635f493fbe4308af1623e8ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f20ad94-5a", "ovs_interfaceid": "0f20ad94-5a78-40ab-9ce7-1a56c05d20bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.253374] env[62368]: DEBUG oslo_concurrency.lockutils [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] Acquired lock "refresh_cache-d865d433-2341-4f6c-b840-609530ae2e51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.253587] env[62368]: DEBUG nova.network.neutron [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Refreshing network info cache for port 0f20ad94-5a78-40ab-9ce7-1a56c05d20bc {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 813.254992] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:b6:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f44b2fa3-6730-4b87-8839-947eff21213f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0f20ad94-5a78-40ab-9ce7-1a56c05d20bc', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.263442] env[62368]: DEBUG oslo.service.loopingcall [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.270572] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 813.273330] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66ec449f-9fb9-4243-a224-2abc09677318 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.288472] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1cb6d0ab-28f5-444a-b9c3-2cd16a8c95ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.290530] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b537e9a0-9ccb-4732-9a4c-110304b3f299 tempest-SecurityGroupsTestJSON-113662258 tempest-SecurityGroupsTestJSON-113662258-project-member] Lock "0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.146s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.307684] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.307684] env[62368]: value = "task-1198475" [ 813.307684] env[62368]: _type = "Task" [ 813.307684] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.321307] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198475, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.446177] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.456725] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198474, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.531219] env[62368]: DEBUG oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a4e4ac-5c09-022c-3adc-c4b8e1904665/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 813.531219] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec55cac-7789-4328-bb16-efc400102d4d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.548255] env[62368]: DEBUG oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a4e4ac-5c09-022c-3adc-c4b8e1904665/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 813.548950] env[62368]: ERROR oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a4e4ac-5c09-022c-3adc-c4b8e1904665/disk-0.vmdk due to incomplete transfer. [ 813.548950] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-be763697-94bf-4505-bebc-4e9793b2f9b7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.565264] env[62368]: DEBUG oslo_vmware.rw_handles [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a4e4ac-5c09-022c-3adc-c4b8e1904665/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 813.565480] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Uploaded image 3b178833-ecf1-4a1f-8ffe-ee20324be047 to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 813.568515] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 813.570083] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-af0fdf45-54d7-47b4-87a1-cf2a0f14992e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.578623] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 813.578623] env[62368]: value = "task-1198476" [ 813.578623] env[62368]: _type = "Task" [ 813.578623] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.588552] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198476, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.745884] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7461be-d926-44b9-bba9-1bd5bdd9dac6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.757335] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239ae7fa-79fd-4fe6-afcc-175e05eb6805 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.799134] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc05bff-2a5f-4837-ad78-30f11505b949 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.823561] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedbf218-6fe4-41a5-832c-588f185eab9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.831611] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198475, 'name': CreateVM_Task, 'duration_secs': 0.389618} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.834656] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 813.835503] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.836221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.836221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 813.836627] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6963bd5-882e-4dea-9b74-add1fa06c726 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.848824] env[62368]: DEBUG nova.compute.provider_tree [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.857599] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 813.857599] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]525da4fc-c6a4-ce93-4ae4-fd3a2ddf4851" [ 813.857599] env[62368]: _type = "Task" [ 813.857599] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.869802] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525da4fc-c6a4-ce93-4ae4-fd3a2ddf4851, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.952392] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198474, 'name': Rename_Task, 'duration_secs': 0.622054} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.953024] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 813.953680] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-197888e5-c309-4df9-b091-80243185e1e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.964803] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 813.964803] env[62368]: value = "task-1198477" [ 813.964803] env[62368]: _type = "Task" [ 813.964803] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.983600] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.089159] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Completed reading data from the image iterator. {{(pid=62368) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 814.089159] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520adb14-6d7a-5cbd-4ea1-f1b8c4b8280b/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 814.093243] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f0689d-9850-4046-93fe-e5dfc5509836 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.101697] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198476, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.106838] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520adb14-6d7a-5cbd-4ea1-f1b8c4b8280b/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 814.107064] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520adb14-6d7a-5cbd-4ea1-f1b8c4b8280b/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 814.107307] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a6bbd9fa-ea80-4730-9392-bc1cf54f026e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.295200] env[62368]: DEBUG nova.network.neutron [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Updated VIF entry in instance network info cache for port 0f20ad94-5a78-40ab-9ce7-1a56c05d20bc. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 814.297122] env[62368]: DEBUG nova.network.neutron [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Updating instance_info_cache with network_info: [{"id": "0f20ad94-5a78-40ab-9ce7-1a56c05d20bc", "address": "fa:16:3e:08:b6:3a", "network": {"id": "e89b0140-71be-4766-9ec4-c37da482a0e3", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2110723737-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb560320635f493fbe4308af1623e8ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f20ad94-5a", "ovs_interfaceid": "0f20ad94-5a78-40ab-9ce7-1a56c05d20bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.342879] env[62368]: DEBUG oslo_vmware.rw_handles [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520adb14-6d7a-5cbd-4ea1-f1b8c4b8280b/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 814.343189] env[62368]: INFO nova.virt.vmwareapi.images [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Downloaded image file data 6281ea63-c477-4e65-8fac-7b43765b0423 [ 814.344381] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea3c3c3-0c35-4fb6-941d-d812e5a9d498 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.361675] env[62368]: DEBUG nova.scheduler.client.report [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.368522] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca4cbe89-b3d4-4ed0-80f7-71483a56be76 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.381140] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525da4fc-c6a4-ce93-4ae4-fd3a2ddf4851, 'name': SearchDatastore_Task, 'duration_secs': 0.0135} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.381140] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.381140] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.381140] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.381140] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.381140] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.381140] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85047eef-618a-4276-b4e1-6a45cbc0650f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.391107] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.391107] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 814.391486] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93e22595-e968-45a8-a0a1-87c9afd9fd44 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.396694] env[62368]: INFO nova.virt.vmwareapi.images [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] The imported VM was unregistered [ 814.398521] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Caching image {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 814.398766] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating directory with path [datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.399817] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f37650c5-1b83-4ef4-9d27-a63ee39361f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.404043] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 814.404043] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52b834c1-32aa-6c2e-7d52-c37fec9904a7" [ 814.404043] env[62368]: _type = "Task" [ 814.404043] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.414404] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b834c1-32aa-6c2e-7d52-c37fec9904a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.415575] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Created directory with path [datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.415761] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0/OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0.vmdk to [datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423/6281ea63-c477-4e65-8fac-7b43765b0423.vmdk. {{(pid=62368) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 814.416023] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f7a9772d-364b-44d4-a55a-e2a9241f8f3a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.423754] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 814.423754] env[62368]: value = "task-1198479" [ 814.423754] env[62368]: _type = "Task" [ 814.423754] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.435711] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198479, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.482332] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198477, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.542499] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.542751] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.594474] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198476, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.798561] env[62368]: DEBUG oslo_concurrency.lockutils [req-028a209d-7849-4afe-9e7f-40a7901281e3 req-f436d39c-c1fc-4821-81a5-e5156ae6bf3c service nova] Releasing lock "refresh_cache-d865d433-2341-4f6c-b840-609530ae2e51" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.871699] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.133s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.874712] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.454s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.877603] env[62368]: INFO nova.compute.claims [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.924339] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b834c1-32aa-6c2e-7d52-c37fec9904a7, 'name': SearchDatastore_Task, 'duration_secs': 0.010731} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.929408] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a486cd97-14c7-41c6-9d87-4323628e1e67 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.940325] env[62368]: INFO nova.scheduler.client.report [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Deleted allocations for instance 300c6c32-b365-40c2-bc7d-6ce028f10476 [ 814.954459] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198479, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.954815] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 814.954815] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5267d446-96f3-ad7e-597e-1e9cc18e3986" [ 814.954815] env[62368]: _type = "Task" [ 814.954815] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.973160] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5267d446-96f3-ad7e-597e-1e9cc18e3986, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.984623] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198477, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.095362] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198476, 'name': Destroy_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.440256] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198479, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.464207] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b964e273-4de5-4a6a-848a-a4a00b807ddc tempest-ServersTestMultiNic-64025196 tempest-ServersTestMultiNic-64025196-project-member] Lock "300c6c32-b365-40c2-bc7d-6ce028f10476" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.417s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.470842] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5267d446-96f3-ad7e-597e-1e9cc18e3986, 'name': SearchDatastore_Task, 'duration_secs': 0.089116} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.471327] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.471543] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] d865d433-2341-4f6c-b840-609530ae2e51/d865d433-2341-4f6c-b840-609530ae2e51.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 815.471964] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b574087-f241-4d24-a4cb-5f8137cf2c37 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.486058] env[62368]: DEBUG oslo_vmware.api [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198477, 'name': PowerOnVM_Task, 'duration_secs': 1.265188} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.488426] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 815.488726] env[62368]: INFO nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Took 9.88 seconds to spawn the instance on the hypervisor. [ 815.488965] env[62368]: DEBUG nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.489355] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 815.489355] env[62368]: value = "task-1198480" [ 815.489355] env[62368]: _type = "Task" [ 815.489355] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.490478] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8544c61-b62d-490e-ab98-3e70070ba941 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.507594] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.597215] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198476, 'name': Destroy_Task, 'duration_secs': 1.530592} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.597979] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Destroyed the VM [ 815.598289] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 815.598641] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3fd4b9c2-4190-41bc-9516-b288fd79c3c5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.612247] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 815.612247] env[62368]: value = "task-1198481" [ 815.612247] env[62368]: _type = "Task" [ 815.612247] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.625738] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.944508] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198479, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.015155] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.021337] env[62368]: INFO nova.compute.manager [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Took 31.59 seconds to build instance. [ 816.124292] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.303022] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c752cd6f-7b5f-4b3f-a985-cf4f045ceb4e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.317611] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c0309a-924a-4489-b872-27244b85c51e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.355518] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d4fc2e-c027-4373-810b-7340c801c550 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.366698] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdcdb8b4-59cd-4900-8d93-0ea1d0793cf8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.384985] env[62368]: DEBUG nova.compute.provider_tree [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.441929] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198479, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.507151] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.523458] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8d3ad85f-12c9-4730-941b-4097bee8ff47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.301s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.626070] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198481, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.889754] env[62368]: DEBUG nova.scheduler.client.report [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.937471] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198479, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.010928] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198480, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.026610] env[62368]: DEBUG nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 817.124660] env[62368]: DEBUG oslo_vmware.api [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198481, 'name': RemoveSnapshot_Task, 'duration_secs': 1.173169} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.125018] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 817.125343] env[62368]: INFO nova.compute.manager [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Took 20.16 seconds to snapshot the instance on the hypervisor. [ 817.396366] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.396940] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.403446] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.217s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.404857] env[62368]: INFO nova.compute.claims [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.448420] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198479, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.710802} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.451276] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0/OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0.vmdk to [datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423/6281ea63-c477-4e65-8fac-7b43765b0423.vmdk. [ 817.451276] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Cleaning up location [datastore1] OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 817.451276] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_dd090762-6080-49ac-8b67-1ee356c253f0 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 817.456513] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04943b4e-0512-49b0-abd3-5d65e600071f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.463352] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 817.463352] env[62368]: value = "task-1198482" [ 817.463352] env[62368]: _type = "Task" [ 817.463352] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.473681] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.512837] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198480, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.810313} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.513154] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] d865d433-2341-4f6c-b840-609530ae2e51/d865d433-2341-4f6c-b840-609530ae2e51.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 817.513439] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.513731] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf07c2f5-ca93-4f9d-9fa3-eb7808d54acb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.522312] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 817.522312] env[62368]: value = "task-1198483" [ 817.522312] env[62368]: _type = "Task" [ 817.522312] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.536106] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.554174] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.669279] env[62368]: DEBUG nova.compute.manager [req-e3eaa587-0cc8-45ff-bddc-e1f40cfa9a48 req-4c0d4ccb-188e-4e6a-9bee-4fed8dbaa1a9 service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Received event network-changed-46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.671966] env[62368]: DEBUG nova.compute.manager [req-e3eaa587-0cc8-45ff-bddc-e1f40cfa9a48 req-4c0d4ccb-188e-4e6a-9bee-4fed8dbaa1a9 service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Refreshing instance network info cache due to event network-changed-46fba643-185a-4fb2-ab24-dd4350aa5ac4. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.671966] env[62368]: DEBUG oslo_concurrency.lockutils [req-e3eaa587-0cc8-45ff-bddc-e1f40cfa9a48 req-4c0d4ccb-188e-4e6a-9bee-4fed8dbaa1a9 service nova] Acquiring lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.671966] env[62368]: DEBUG oslo_concurrency.lockutils [req-e3eaa587-0cc8-45ff-bddc-e1f40cfa9a48 req-4c0d4ccb-188e-4e6a-9bee-4fed8dbaa1a9 service nova] Acquired lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.671966] env[62368]: DEBUG nova.network.neutron [req-e3eaa587-0cc8-45ff-bddc-e1f40cfa9a48 req-4c0d4ccb-188e-4e6a-9bee-4fed8dbaa1a9 service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Refreshing network info cache for port 46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 817.682219] env[62368]: DEBUG nova.compute.manager [None req-08164ba6-d677-49a9-babd-22546c646a66 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Found 1 images (rotation: 2) {{(pid=62368) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 817.910444] env[62368]: DEBUG nova.compute.utils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.914322] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.918562] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 817.960264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7838d3c-43df-4420-87f5-7104f62363b8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.960264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7838d3c-43df-4420-87f5-7104f62363b8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.960264] env[62368]: DEBUG nova.objects.instance [None req-e7838d3c-43df-4420-87f5-7104f62363b8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'flavor' on Instance uuid 67d8c43a-d3af-407b-847a-1c2c8c53dbfb {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.971055] env[62368]: DEBUG nova.policy [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8895edd603ef41aca1f01c7d268bf8a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1bc06cf02fa4cebb141a45d732b123a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.979516] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.041903} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.979648] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.979794] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423/6281ea63-c477-4e65-8fac-7b43765b0423.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.980523] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423/6281ea63-c477-4e65-8fac-7b43765b0423.vmdk to [datastore1] 0b861a8b-dcad-46ad-90ef-a46450c5cd51/0b861a8b-dcad-46ad-90ef-a46450c5cd51.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 817.980523] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21eab0a6-cd92-4ff0-b755-9270ddbd26d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.991690] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 817.991690] env[62368]: value = "task-1198484" [ 817.991690] env[62368]: _type = "Task" [ 817.991690] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.002061] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198484, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.033968] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.309317] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Successfully created port: cdba9d6e-63cb-4936-885a-9dc035e3dff0 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.427300] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.465337] env[62368]: DEBUG nova.objects.instance [None req-e7838d3c-43df-4420-87f5-7104f62363b8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'pci_requests' on Instance uuid 67d8c43a-d3af-407b-847a-1c2c8c53dbfb {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.511025] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198484, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.542987] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.818784] env[62368]: DEBUG nova.network.neutron [req-e3eaa587-0cc8-45ff-bddc-e1f40cfa9a48 req-4c0d4ccb-188e-4e6a-9bee-4fed8dbaa1a9 service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updated VIF entry in instance network info cache for port 46fba643-185a-4fb2-ab24-dd4350aa5ac4. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 818.819342] env[62368]: DEBUG nova.network.neutron [req-e3eaa587-0cc8-45ff-bddc-e1f40cfa9a48 req-4c0d4ccb-188e-4e6a-9bee-4fed8dbaa1a9 service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updating instance_info_cache with network_info: [{"id": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "address": "fa:16:3e:23:0d:cc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46fba643-18", "ovs_interfaceid": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.896728] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161deb3b-fd27-44d1-a766-e3b19e6bbb02 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.907499] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2853f1c-7899-439d-b512-f900c3505489 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.953637] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f0496f-daa3-4d1c-8ad7-a755d331b161 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.967989] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558d0dee-2d11-47be-b7c2-324eeb4dc640 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.978532] env[62368]: DEBUG nova.objects.base [None req-e7838d3c-43df-4420-87f5-7104f62363b8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Object Instance<67d8c43a-d3af-407b-847a-1c2c8c53dbfb> lazy-loaded attributes: flavor,pci_requests {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 818.978532] env[62368]: DEBUG nova.network.neutron [None req-e7838d3c-43df-4420-87f5-7104f62363b8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 818.997721] env[62368]: DEBUG nova.compute.provider_tree [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.012863] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198484, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.040929] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.189921} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.041390] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.042431] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050b7c56-0829-4e3d-aaa0-031607330c05 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.070994] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] d865d433-2341-4f6c-b840-609530ae2e51/d865d433-2341-4f6c-b840-609530ae2e51.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.070994] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45dc26f0-9b63-420b-858d-54373b97190e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.088988] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7838d3c-43df-4420-87f5-7104f62363b8 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.130s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.100067] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 819.100067] env[62368]: value = "task-1198485" [ 819.100067] env[62368]: _type = "Task" [ 819.100067] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.113072] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198485, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.125247] env[62368]: DEBUG nova.compute.manager [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.126594] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41c4d96-cb20-4cc5-83d8-04a1dbcbee61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.322987] env[62368]: DEBUG oslo_concurrency.lockutils [req-e3eaa587-0cc8-45ff-bddc-e1f40cfa9a48 req-4c0d4ccb-188e-4e6a-9bee-4fed8dbaa1a9 service nova] Releasing lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.458748] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.489872] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.490174] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.490339] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.490525] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.490674] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.490825] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.491047] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.491217] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.491389] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.491758] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.491981] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.492953] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9179d68d-9bfa-4095-8026-5721f2f4b7a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.504327] env[62368]: DEBUG nova.scheduler.client.report [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.510951] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3496976-cecc-47d2-8bc4-3b59850f3d2f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.518887] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198484, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.615714] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.641824] env[62368]: INFO nova.compute.manager [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] instance snapshotting [ 819.642556] env[62368]: DEBUG nova.objects.instance [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'flavor' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.013511] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198484, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.016693] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.017270] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 820.019944] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.115s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.020286] env[62368]: DEBUG nova.objects.instance [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 820.114803] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.151022] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25e15f6-6f81-43d9-92aa-1253b3b6ff61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.175036] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942994dd-c5c8-4caf-b4a2-69c9999e85a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.311206] env[62368]: DEBUG nova.compute.manager [req-6225522c-2811-4ef7-8683-5b5088f38d45 req-6395d0fc-efff-4939-b5f8-72faf122ecbc service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Received event network-vif-plugged-cdba9d6e-63cb-4936-885a-9dc035e3dff0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.311206] env[62368]: DEBUG oslo_concurrency.lockutils [req-6225522c-2811-4ef7-8683-5b5088f38d45 req-6395d0fc-efff-4939-b5f8-72faf122ecbc service nova] Acquiring lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.311206] env[62368]: DEBUG oslo_concurrency.lockutils [req-6225522c-2811-4ef7-8683-5b5088f38d45 req-6395d0fc-efff-4939-b5f8-72faf122ecbc service nova] Lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.311206] env[62368]: DEBUG oslo_concurrency.lockutils [req-6225522c-2811-4ef7-8683-5b5088f38d45 req-6395d0fc-efff-4939-b5f8-72faf122ecbc service nova] Lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.311206] env[62368]: DEBUG nova.compute.manager [req-6225522c-2811-4ef7-8683-5b5088f38d45 req-6395d0fc-efff-4939-b5f8-72faf122ecbc service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] No waiting events found dispatching network-vif-plugged-cdba9d6e-63cb-4936-885a-9dc035e3dff0 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.311206] env[62368]: WARNING nova.compute.manager [req-6225522c-2811-4ef7-8683-5b5088f38d45 req-6395d0fc-efff-4939-b5f8-72faf122ecbc service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Received unexpected event network-vif-plugged-cdba9d6e-63cb-4936-885a-9dc035e3dff0 for instance with vm_state building and task_state spawning. [ 820.396677] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Successfully updated port: cdba9d6e-63cb-4936-885a-9dc035e3dff0 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.510058] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198484, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.524680] env[62368]: DEBUG nova.compute.utils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.529305] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.529485] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 820.610597] env[62368]: DEBUG nova.policy [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8895edd603ef41aca1f01c7d268bf8a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1bc06cf02fa4cebb141a45d732b123a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.619679] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.691650] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 820.691827] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e2895275-eeea-48a7-8c7a-0ba43be6c1d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.706914] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 820.706914] env[62368]: value = "task-1198486" [ 820.706914] env[62368]: _type = "Task" [ 820.706914] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.717677] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198486, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.899836] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "refresh_cache-29d2f0d6-7a76-4039-acbb-a3abd69d8370" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.900020] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "refresh_cache-29d2f0d6-7a76-4039-acbb-a3abd69d8370" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.900209] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.011944] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198484, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.551908} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.011944] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6281ea63-c477-4e65-8fac-7b43765b0423/6281ea63-c477-4e65-8fac-7b43765b0423.vmdk to [datastore1] 0b861a8b-dcad-46ad-90ef-a46450c5cd51/0b861a8b-dcad-46ad-90ef-a46450c5cd51.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 821.012318] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce09eb21-a86f-4ace-9fbc-bc42c73a46c7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.034077] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 821.038463] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b4dbd19e-4781-490f-bed3-4dcc8dfe3388 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.050374] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 0b861a8b-dcad-46ad-90ef-a46450c5cd51/0b861a8b-dcad-46ad-90ef-a46450c5cd51.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.051317] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.737s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.052300] env[62368]: DEBUG nova.objects.instance [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lazy-loading 'resources' on Instance uuid 0f1bff84-1fb9-471a-b685-7d527b376b0a {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.054189] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e9a5f57-0e18-42db-8571-598527f8fc88 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.070858] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Successfully created port: 07be57e1-db65-4ba2-a054-72ee59275498 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.080881] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 821.080881] env[62368]: value = "task-1198487" [ 821.080881] env[62368]: _type = "Task" [ 821.080881] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.092384] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198487, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.115724] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198485, 'name': ReconfigVM_Task, 'duration_secs': 1.636745} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.116161] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Reconfigured VM instance instance-00000043 to attach disk [datastore1] d865d433-2341-4f6c-b840-609530ae2e51/d865d433-2341-4f6c-b840-609530ae2e51.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.116895] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f99d5060-971b-4f02-8bce-021725953cbc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.127019] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.127019] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.127019] env[62368]: DEBUG nova.objects.instance [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'flavor' on Instance uuid 67d8c43a-d3af-407b-847a-1c2c8c53dbfb {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.128619] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 821.128619] env[62368]: value = "task-1198488" [ 821.128619] env[62368]: _type = "Task" [ 821.128619] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.139524] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198488, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.220851] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198486, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.465139] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.595712] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198487, 'name': ReconfigVM_Task, 'duration_secs': 0.321567} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.601021] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 0b861a8b-dcad-46ad-90ef-a46450c5cd51/0b861a8b-dcad-46ad-90ef-a46450c5cd51.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.601021] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8a93321-de93-4bd3-a4d5-8e049bc47055 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.610426] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 821.610426] env[62368]: value = "task-1198489" [ 821.610426] env[62368]: _type = "Task" [ 821.610426] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.625060] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198489, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.647820] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198488, 'name': Rename_Task, 'duration_secs': 0.164665} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.647942] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 821.648141] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5069b6f7-1cef-4c78-bef2-9cf7a8c03d0b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.657454] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 821.657454] env[62368]: value = "task-1198490" [ 821.657454] env[62368]: _type = "Task" [ 821.657454] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.669118] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.716676] env[62368]: DEBUG nova.objects.instance [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'pci_requests' on Instance uuid 67d8c43a-d3af-407b-847a-1c2c8c53dbfb {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.722558] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198486, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.949620] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdb53d2-e5ca-449b-b616-2a5adb1ad906 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.958214] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Updating instance_info_cache with network_info: [{"id": "cdba9d6e-63cb-4936-885a-9dc035e3dff0", "address": "fa:16:3e:33:b6:85", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdba9d6e-63", "ovs_interfaceid": "cdba9d6e-63cb-4936-885a-9dc035e3dff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.963715] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e093d4ff-f8cb-4c32-a007-e23978f0bebb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.001219] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e3d9e0-0f9d-4819-9fee-4752ed808dc9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.010668] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32235346-9b18-4127-b86a-0bf6b3f53258 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.025785] env[62368]: DEBUG nova.compute.provider_tree [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.058276] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 822.089753] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.089876] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.090111] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.090414] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.090581] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.090810] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.091116] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.091350] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.091624] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.091846] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.092130] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.093363] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466f5305-12c5-4a61-8fc9-769c0f01a279 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.104059] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1caf9d-b986-434d-b2d6-da240bec21e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.138201] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198489, 'name': Rename_Task, 'duration_secs': 0.163931} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.138463] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.138865] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ab6c580-ee7b-4539-b0df-85adb73be831 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.154263] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 822.154263] env[62368]: value = "task-1198491" [ 822.154263] env[62368]: _type = "Task" [ 822.154263] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.170014] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198491, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.178322] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198490, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.219366] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198486, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.225338] env[62368]: DEBUG nova.objects.base [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Object Instance<67d8c43a-d3af-407b-847a-1c2c8c53dbfb> lazy-loaded attributes: flavor,pci_requests {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 822.225338] env[62368]: DEBUG nova.network.neutron [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 822.344176] env[62368]: DEBUG nova.policy [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68c2209271e043b698fd3794955f40bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcecae6c09124e9fa4dcd5d63833b0bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.464405] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "refresh_cache-29d2f0d6-7a76-4039-acbb-a3abd69d8370" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.464701] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Instance network_info: |[{"id": "cdba9d6e-63cb-4936-885a-9dc035e3dff0", "address": "fa:16:3e:33:b6:85", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdba9d6e-63", "ovs_interfaceid": "cdba9d6e-63cb-4936-885a-9dc035e3dff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.466250] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:b6:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13b62154-a0e1-4eed-bc30-6464b15993bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdba9d6e-63cb-4936-885a-9dc035e3dff0', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.475257] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Creating folder: Project (e1bc06cf02fa4cebb141a45d732b123a). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 822.475611] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd494914-2ffc-4d37-9454-8eea2135f597 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.491140] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Created folder: Project (e1bc06cf02fa4cebb141a45d732b123a) in parent group-v259706. [ 822.491384] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Creating folder: Instances. Parent ref: group-v259807. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 822.491701] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1cac6b2-bec2-416d-9bc9-b5a5dd9b070f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.503426] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Created folder: Instances in parent group-v259807. [ 822.503426] env[62368]: DEBUG oslo.service.loopingcall [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.503759] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 822.503759] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d28f3a3-9c14-44d7-a89e-439f77b60bbb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.528020] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.528020] env[62368]: value = "task-1198494" [ 822.528020] env[62368]: _type = "Task" [ 822.528020] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.531945] env[62368]: DEBUG nova.scheduler.client.report [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.543238] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198494, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.661534] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198491, 'name': PowerOnVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.671296] env[62368]: DEBUG oslo_vmware.api [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198490, 'name': PowerOnVM_Task, 'duration_secs': 0.547983} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.671614] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 822.672194] env[62368]: INFO nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Took 11.94 seconds to spawn the instance on the hypervisor. [ 822.672293] env[62368]: DEBUG nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.673380] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755b2a75-c80c-4078-8cc3-15395107309f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.678106] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Successfully updated port: 07be57e1-db65-4ba2-a054-72ee59275498 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.725754] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198486, 'name': CreateSnapshot_Task, 'duration_secs': 1.557814} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.726384] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 822.728469] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260a0d34-dfa6-4837-a64a-73399565bc5a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.940438] env[62368]: DEBUG nova.network.neutron [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Successfully created port: 90ff566e-24b4-4a83-a965-0bc25b1152e0 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.999846] env[62368]: DEBUG nova.compute.manager [req-1a2ac193-c69f-449c-9906-ad032b561be4 req-8a0f3d8f-3c6b-4cd0-93ee-4db8aa0e7c3b service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Received event network-changed-cdba9d6e-63cb-4936-885a-9dc035e3dff0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.000047] env[62368]: DEBUG nova.compute.manager [req-1a2ac193-c69f-449c-9906-ad032b561be4 req-8a0f3d8f-3c6b-4cd0-93ee-4db8aa0e7c3b service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Refreshing instance network info cache due to event network-changed-cdba9d6e-63cb-4936-885a-9dc035e3dff0. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.000279] env[62368]: DEBUG oslo_concurrency.lockutils [req-1a2ac193-c69f-449c-9906-ad032b561be4 req-8a0f3d8f-3c6b-4cd0-93ee-4db8aa0e7c3b service nova] Acquiring lock "refresh_cache-29d2f0d6-7a76-4039-acbb-a3abd69d8370" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.000423] env[62368]: DEBUG oslo_concurrency.lockutils [req-1a2ac193-c69f-449c-9906-ad032b561be4 req-8a0f3d8f-3c6b-4cd0-93ee-4db8aa0e7c3b service nova] Acquired lock "refresh_cache-29d2f0d6-7a76-4039-acbb-a3abd69d8370" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.000583] env[62368]: DEBUG nova.network.neutron [req-1a2ac193-c69f-449c-9906-ad032b561be4 req-8a0f3d8f-3c6b-4cd0-93ee-4db8aa0e7c3b service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Refreshing network info cache for port cdba9d6e-63cb-4936-885a-9dc035e3dff0 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.040114] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.040582] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198494, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.041971] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.058s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.046871] env[62368]: INFO nova.compute.claims [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.071963] env[62368]: INFO nova.scheduler.client.report [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Deleted allocations for instance 0f1bff84-1fb9-471a-b685-7d527b376b0a [ 823.161403] env[62368]: DEBUG oslo_vmware.api [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198491, 'name': PowerOnVM_Task, 'duration_secs': 0.60416} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.161801] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.162107] env[62368]: INFO nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Took 15.02 seconds to spawn the instance on the hypervisor. [ 823.162388] env[62368]: DEBUG nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.163416] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fe96fd-ed56-4348-b2e2-15eedee2721a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.186108] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "refresh_cache-cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.186593] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "refresh_cache-cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.186872] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.201372] env[62368]: INFO nova.compute.manager [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Took 37.76 seconds to build instance. [ 823.252752] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 823.253491] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-90711525-d3c4-4a90-afb9-5da3d0e1a505 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.264308] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 823.264308] env[62368]: value = "task-1198495" [ 823.264308] env[62368]: _type = "Task" [ 823.264308] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.274702] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198495, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.540657] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198494, 'name': CreateVM_Task, 'duration_secs': 0.536736} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.541159] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 823.542162] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.542346] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.542771] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.542953] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4078570-a8a4-4604-9111-d7af09c51c00 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.554506] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 823.554506] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5209d341-5a60-cd3f-9c79-c48acd7afe86" [ 823.554506] env[62368]: _type = "Task" [ 823.554506] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.566486] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5209d341-5a60-cd3f-9c79-c48acd7afe86, 'name': SearchDatastore_Task, 'duration_secs': 0.012966} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.567774] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.568102] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.568296] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.568659] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.569166] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.569503] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa84878f-b6ac-499a-8612-600d1b9c197f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.579762] env[62368]: DEBUG oslo_concurrency.lockutils [None req-572a5704-aa72-4eba-8618-d2e3d242eab8 tempest-AttachInterfacesV270Test-725696446 tempest-AttachInterfacesV270Test-725696446-project-member] Lock "0f1bff84-1fb9-471a-b685-7d527b376b0a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.707s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.582482] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.582811] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 823.583819] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f70f080c-ee21-4e46-a771-3185f929d44a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.590954] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 823.590954] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5247ab47-90c9-4f13-31d2-3701a25a0185" [ 823.590954] env[62368]: _type = "Task" [ 823.590954] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.603303] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5247ab47-90c9-4f13-31d2-3701a25a0185, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.685990] env[62368]: INFO nova.compute.manager [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Took 38.41 seconds to build instance. [ 823.707382] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b99198aa-908c-4e98-8e61-42e750bd65d0 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "d865d433-2341-4f6c-b840-609530ae2e51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.981s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.752203] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.782405] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198495, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.070876] env[62368]: DEBUG nova.network.neutron [req-1a2ac193-c69f-449c-9906-ad032b561be4 req-8a0f3d8f-3c6b-4cd0-93ee-4db8aa0e7c3b service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Updated VIF entry in instance network info cache for port cdba9d6e-63cb-4936-885a-9dc035e3dff0. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 824.071385] env[62368]: DEBUG nova.network.neutron [req-1a2ac193-c69f-449c-9906-ad032b561be4 req-8a0f3d8f-3c6b-4cd0-93ee-4db8aa0e7c3b service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Updating instance_info_cache with network_info: [{"id": "cdba9d6e-63cb-4936-885a-9dc035e3dff0", "address": "fa:16:3e:33:b6:85", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdba9d6e-63", "ovs_interfaceid": "cdba9d6e-63cb-4936-885a-9dc035e3dff0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.110894] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5247ab47-90c9-4f13-31d2-3701a25a0185, 'name': SearchDatastore_Task, 'duration_secs': 0.014426} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.110894] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fbf8856-d31d-40f6-9fe7-6acb26f3dace {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.122338] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 824.122338] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52b6a554-2dc8-84d2-1821-e3ba62f35e55" [ 824.122338] env[62368]: _type = "Task" [ 824.122338] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.134646] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b6a554-2dc8-84d2-1821-e3ba62f35e55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.190187] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5a2eb96-d70d-45ef-b13e-bc2a3a747cd3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.878s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.196527] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Updating instance_info_cache with network_info: [{"id": "07be57e1-db65-4ba2-a054-72ee59275498", "address": "fa:16:3e:0c:43:ff", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07be57e1-db", "ovs_interfaceid": "07be57e1-db65-4ba2-a054-72ee59275498", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.211672] env[62368]: DEBUG nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.278534] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198495, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.534198] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab68c61-a4c6-4fe9-b941-31c66d37549f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.545165] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cf52d8-973b-48f0-96b4-bef7d8ad111b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.583218] env[62368]: DEBUG oslo_concurrency.lockutils [req-1a2ac193-c69f-449c-9906-ad032b561be4 req-8a0f3d8f-3c6b-4cd0-93ee-4db8aa0e7c3b service nova] Releasing lock "refresh_cache-29d2f0d6-7a76-4039-acbb-a3abd69d8370" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.585392] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0b6e66-cd77-4e05-8be3-5e052c1735aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.593901] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "0cb0097b-2908-4107-bb57-0241915ccff7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.594186] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "0cb0097b-2908-4107-bb57-0241915ccff7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.599601] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130358d3-9664-4a90-a1d3-144118b93812 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.618270] env[62368]: DEBUG nova.compute.provider_tree [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.633623] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b6a554-2dc8-84d2-1821-e3ba62f35e55, 'name': SearchDatastore_Task, 'duration_secs': 0.011749} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.634382] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.634382] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 29d2f0d6-7a76-4039-acbb-a3abd69d8370/29d2f0d6-7a76-4039-acbb-a3abd69d8370.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 824.634933] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8013626-fb19-4b44-9aef-db01339052f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.643961] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 824.643961] env[62368]: value = "task-1198496" [ 824.643961] env[62368]: _type = "Task" [ 824.643961] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.655691] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198496, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.663036] env[62368]: DEBUG nova.network.neutron [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Successfully updated port: 90ff566e-24b4-4a83-a965-0bc25b1152e0 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.703832] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "refresh_cache-cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.703832] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Instance network_info: |[{"id": "07be57e1-db65-4ba2-a054-72ee59275498", "address": "fa:16:3e:0c:43:ff", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07be57e1-db", "ovs_interfaceid": "07be57e1-db65-4ba2-a054-72ee59275498", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.704408] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:43:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13b62154-a0e1-4eed-bc30-6464b15993bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07be57e1-db65-4ba2-a054-72ee59275498', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.712439] env[62368]: DEBUG oslo.service.loopingcall [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.712998] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 824.713430] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b08d82e-f467-4254-8415-59441409837c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.739407] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.739407] env[62368]: value = "task-1198497" [ 824.739407] env[62368]: _type = "Task" [ 824.739407] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.748380] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.755899] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198497, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.779556] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198495, 'name': CloneVM_Task, 'duration_secs': 1.391835} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.783019] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Created linked-clone VM from snapshot [ 824.783019] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98378e4e-159c-4d75-89cb-5d3067eeb8e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.789893] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Uploading image ba6b39d1-f7e6-4aec-93f4-f648afffeff8 {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 824.826814] env[62368]: DEBUG oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 824.826814] env[62368]: value = "vm-259810" [ 824.826814] env[62368]: _type = "VirtualMachine" [ 824.826814] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 824.827194] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ef950f6e-ad69-4458-ac6a-80e0fe382f1f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.836877] env[62368]: DEBUG oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease: (returnval){ [ 824.836877] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520f8c99-4ac1-ef75-6f03-d0da43096f87" [ 824.836877] env[62368]: _type = "HttpNfcLease" [ 824.836877] env[62368]: } obtained for exporting VM: (result){ [ 824.836877] env[62368]: value = "vm-259810" [ 824.836877] env[62368]: _type = "VirtualMachine" [ 824.836877] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 824.837223] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the lease: (returnval){ [ 824.837223] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520f8c99-4ac1-ef75-6f03-d0da43096f87" [ 824.837223] env[62368]: _type = "HttpNfcLease" [ 824.837223] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 824.844536] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 824.844536] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520f8c99-4ac1-ef75-6f03-d0da43096f87" [ 824.844536] env[62368]: _type = "HttpNfcLease" [ 824.844536] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 825.097869] env[62368]: DEBUG nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.122990] env[62368]: DEBUG nova.scheduler.client.report [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.163527] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198496, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.167044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.167044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.167509] env[62368]: DEBUG nova.network.neutron [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 825.195781] env[62368]: DEBUG nova.compute.manager [req-ce4878a9-eaf8-43b5-8942-1c0533bc1979 req-e55185dc-6e99-4d6a-81b8-ec5d055b865d service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-vif-plugged-90ff566e-24b4-4a83-a965-0bc25b1152e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.195781] env[62368]: DEBUG oslo_concurrency.lockutils [req-ce4878a9-eaf8-43b5-8942-1c0533bc1979 req-e55185dc-6e99-4d6a-81b8-ec5d055b865d service nova] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.195893] env[62368]: DEBUG oslo_concurrency.lockutils [req-ce4878a9-eaf8-43b5-8942-1c0533bc1979 req-e55185dc-6e99-4d6a-81b8-ec5d055b865d service nova] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.196365] env[62368]: DEBUG oslo_concurrency.lockutils [req-ce4878a9-eaf8-43b5-8942-1c0533bc1979 req-e55185dc-6e99-4d6a-81b8-ec5d055b865d service nova] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.196651] env[62368]: DEBUG nova.compute.manager [req-ce4878a9-eaf8-43b5-8942-1c0533bc1979 req-e55185dc-6e99-4d6a-81b8-ec5d055b865d service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] No waiting events found dispatching network-vif-plugged-90ff566e-24b4-4a83-a965-0bc25b1152e0 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.196947] env[62368]: WARNING nova.compute.manager [req-ce4878a9-eaf8-43b5-8942-1c0533bc1979 req-e55185dc-6e99-4d6a-81b8-ec5d055b865d service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received unexpected event network-vif-plugged-90ff566e-24b4-4a83-a965-0bc25b1152e0 for instance with vm_state active and task_state None. [ 825.251293] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198497, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.294785] env[62368]: DEBUG nova.compute.manager [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Received event network-vif-plugged-07be57e1-db65-4ba2-a054-72ee59275498 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.294986] env[62368]: DEBUG oslo_concurrency.lockutils [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] Acquiring lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.295261] env[62368]: DEBUG oslo_concurrency.lockutils [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] Lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.295438] env[62368]: DEBUG oslo_concurrency.lockutils [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] Lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.295612] env[62368]: DEBUG nova.compute.manager [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] No waiting events found dispatching network-vif-plugged-07be57e1-db65-4ba2-a054-72ee59275498 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.295778] env[62368]: WARNING nova.compute.manager [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Received unexpected event network-vif-plugged-07be57e1-db65-4ba2-a054-72ee59275498 for instance with vm_state building and task_state spawning. [ 825.295977] env[62368]: DEBUG nova.compute.manager [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Received event network-changed-07be57e1-db65-4ba2-a054-72ee59275498 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.296103] env[62368]: DEBUG nova.compute.manager [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Refreshing instance network info cache due to event network-changed-07be57e1-db65-4ba2-a054-72ee59275498. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.296294] env[62368]: DEBUG oslo_concurrency.lockutils [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] Acquiring lock "refresh_cache-cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.296434] env[62368]: DEBUG oslo_concurrency.lockutils [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] Acquired lock "refresh_cache-cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.296597] env[62368]: DEBUG nova.network.neutron [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Refreshing network info cache for port 07be57e1-db65-4ba2-a054-72ee59275498 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 825.346377] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 825.346377] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520f8c99-4ac1-ef75-6f03-d0da43096f87" [ 825.346377] env[62368]: _type = "HttpNfcLease" [ 825.346377] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 825.346738] env[62368]: DEBUG oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 825.346738] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520f8c99-4ac1-ef75-6f03-d0da43096f87" [ 825.346738] env[62368]: _type = "HttpNfcLease" [ 825.346738] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 825.347477] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf8ce77-bfaf-4d6c-b103-c6e51715ee9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.356046] env[62368]: DEBUG oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d7488d-1323-8742-5da7-9e3c7fdeea0c/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 825.356252] env[62368]: DEBUG oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d7488d-1323-8742-5da7-9e3c7fdeea0c/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 825.495680] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e2952cd4-1aff-40de-b3a5-232e4fb91a40 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.626385] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.628321] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.628984] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.631654] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.726s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.632059] env[62368]: DEBUG nova.objects.instance [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lazy-loading 'resources' on Instance uuid 3f1af54b-392f-432a-9ffa-a133da428f94 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.656938] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198496, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575099} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.657285] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 29d2f0d6-7a76-4039-acbb-a3abd69d8370/29d2f0d6-7a76-4039-acbb-a3abd69d8370.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 825.657673] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.657958] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d87005c7-c26e-4e2a-ac06-b7ff56f9bb78 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.668485] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 825.668485] env[62368]: value = "task-1198499" [ 825.668485] env[62368]: _type = "Task" [ 825.668485] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.682109] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198499, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.709927] env[62368]: WARNING nova.network.neutron [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] 3ee164de-30af-4404-935a-a7f988d824b1 already exists in list: networks containing: ['3ee164de-30af-4404-935a-a7f988d824b1']. ignoring it [ 825.753374] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198497, 'name': CreateVM_Task, 'duration_secs': 0.681903} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.753586] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 825.754909] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.755099] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.755715] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.756635] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d83db057-c527-4e3e-aae4-43ed7829c768 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.764879] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 825.764879] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521f2c79-ed2a-31d3-c65e-30d0285babc6" [ 825.764879] env[62368]: _type = "Task" [ 825.764879] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.771578] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.771885] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.772414] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.772669] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.772841] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.777557] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521f2c79-ed2a-31d3-c65e-30d0285babc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.778327] env[62368]: INFO nova.compute.manager [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Terminating instance [ 825.780629] env[62368]: DEBUG nova.compute.manager [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.780864] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.781809] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82b8daa-e671-4ed8-a108-9d66229ecfa6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.792681] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.792681] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e2f1075-4f64-402f-8c03-c753e8852710 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.806973] env[62368]: DEBUG oslo_vmware.api [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 825.806973] env[62368]: value = "task-1198500" [ 825.806973] env[62368]: _type = "Task" [ 825.806973] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.819146] env[62368]: DEBUG oslo_vmware.api [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.141534] env[62368]: DEBUG nova.compute.utils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.147489] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.147631] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.187401] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198499, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078332} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.187901] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.188722] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376b21ce-ec8d-49ca-b5c1-a82252cac0fb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.194104] env[62368]: DEBUG nova.network.neutron [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Updated VIF entry in instance network info cache for port 07be57e1-db65-4ba2-a054-72ee59275498. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 826.194514] env[62368]: DEBUG nova.network.neutron [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Updating instance_info_cache with network_info: [{"id": "07be57e1-db65-4ba2-a054-72ee59275498", "address": "fa:16:3e:0c:43:ff", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07be57e1-db", "ovs_interfaceid": "07be57e1-db65-4ba2-a054-72ee59275498", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.226722] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 29d2f0d6-7a76-4039-acbb-a3abd69d8370/29d2f0d6-7a76-4039-acbb-a3abd69d8370.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.231180] env[62368]: DEBUG nova.policy [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8895edd603ef41aca1f01c7d268bf8a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1bc06cf02fa4cebb141a45d732b123a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.233804] env[62368]: DEBUG nova.network.neutron [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90ff566e-24b4-4a83-a965-0bc25b1152e0", "address": "fa:16:3e:29:10:46", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ff566e-24", "ovs_interfaceid": "90ff566e-24b4-4a83-a965-0bc25b1152e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.237121] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01c4bb7c-191e-462d-907f-d04da02ac1db {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.253801] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.254580] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.254754] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.258775] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b1f4ef-dd1b-4fe8-9f3c-daf41126d5ef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.284945] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.285332] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.285547] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.285866] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.286262] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.286437] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.287313] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.287497] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.287714] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.287892] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.288212] env[62368]: DEBUG nova.virt.hardware [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.296109] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Reconfiguring VM to attach interface {{(pid=62368) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 826.302212] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8df9054-f5e4-4df9-8ace-0b17ee5b6017 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.317917] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 826.317917] env[62368]: value = "task-1198501" [ 826.317917] env[62368]: _type = "Task" [ 826.317917] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.329066] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521f2c79-ed2a-31d3-c65e-30d0285babc6, 'name': SearchDatastore_Task, 'duration_secs': 0.021581} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.331294] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.331632] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.331903] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.332187] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.332459] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.332950] env[62368]: DEBUG oslo_vmware.api [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 826.332950] env[62368]: value = "task-1198502" [ 826.332950] env[62368]: _type = "Task" [ 826.332950] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.336064] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4e3ef8ec-a6ba-457d-b1ec-f32334834d19 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.347708] env[62368]: DEBUG oslo_vmware.api [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198500, 'name': PowerOffVM_Task, 'duration_secs': 0.229953} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.348345] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198501, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.349270] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 826.349528] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 826.350120] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56b8662b-0556-4d85-98b7-980f26b3cfc2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.358282] env[62368]: DEBUG oslo_vmware.api [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198502, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.359513] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.361021] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 826.361021] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-396b1d71-d835-4051-9679-e382a1c825bc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.368586] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 826.368586] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5204307c-f541-cb99-e8ee-5833c5325ae8" [ 826.368586] env[62368]: _type = "Task" [ 826.368586] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.382234] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5204307c-f541-cb99-e8ee-5833c5325ae8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.438068] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 826.438443] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 826.438702] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleting the datastore file [datastore1] 0b861a8b-dcad-46ad-90ef-a46450c5cd51 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 826.438996] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4a632d0-9030-4ed5-9a20-6a77b74802d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.449238] env[62368]: DEBUG oslo_vmware.api [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 826.449238] env[62368]: value = "task-1198504" [ 826.449238] env[62368]: _type = "Task" [ 826.449238] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.460535] env[62368]: DEBUG oslo_vmware.api [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198504, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.612093] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Successfully created port: b65f0e56-58ab-48a2-90ca-ca285f331170 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.648617] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.698838] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c96391-0926-4aa8-a676-f6f6afcd3a80 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.704818] env[62368]: DEBUG oslo_concurrency.lockutils [req-7347e931-c54d-4500-9785-afc879a2cc16 req-47c9e659-0de1-419e-b5ef-4fe137b92408 service nova] Releasing lock "refresh_cache-cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.711475] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67729160-7ad0-49f5-925b-30017a7d0cbc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.744642] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192c1775-268d-4fe9-9c16-db11b40457a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.754434] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282ca047-20bb-451e-ba62-68c50f992b70 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.759592] env[62368]: DEBUG nova.compute.manager [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.760411] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a5a26f-7d0c-419f-b89f-ec542f02074e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.778733] env[62368]: DEBUG nova.compute.provider_tree [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.803225] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.804026] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.804026] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 826.804183] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Rebuilding the list of instances to heal {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 826.828835] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198501, 'name': ReconfigVM_Task, 'duration_secs': 0.387238} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.829185] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 29d2f0d6-7a76-4039-acbb-a3abd69d8370/29d2f0d6-7a76-4039-acbb-a3abd69d8370.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.829913] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98d4b921-6571-4eb5-86f3-8d88f099b112 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.838968] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 826.838968] env[62368]: value = "task-1198505" [ 826.838968] env[62368]: _type = "Task" [ 826.838968] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.856040] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198505, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.856817] env[62368]: DEBUG oslo_vmware.api [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198502, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.879948] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5204307c-f541-cb99-e8ee-5833c5325ae8, 'name': SearchDatastore_Task, 'duration_secs': 0.015876} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.881096] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80205e89-d267-4be2-bbff-0dca22ea92b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.887971] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 826.887971] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52cfb97c-28ae-f2b9-1872-12d7edbaff07" [ 826.887971] env[62368]: _type = "Task" [ 826.887971] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.898976] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cfb97c-28ae-f2b9-1872-12d7edbaff07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.960575] env[62368]: DEBUG oslo_vmware.api [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198504, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323151} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.960962] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.961227] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 826.961474] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 826.961769] env[62368]: INFO nova.compute.manager [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Took 1.18 seconds to destroy the instance on the hypervisor. [ 826.962048] env[62368]: DEBUG oslo.service.loopingcall [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.962288] env[62368]: DEBUG nova.compute.manager [-] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.962388] env[62368]: DEBUG nova.network.neutron [-] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 827.282028] env[62368]: INFO nova.compute.manager [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] instance snapshotting [ 827.284726] env[62368]: DEBUG nova.scheduler.client.report [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.288968] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73d8c6f-7151-4b03-b982-92c81cc59e70 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.313178] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Skipping network cache update for instance because it is being deleted. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 827.313522] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 827.313637] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 827.313719] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 827.313922] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.314127] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquired lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.314305] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Forcefully refreshing network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 827.314476] env[62368]: DEBUG nova.objects.instance [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lazy-loading 'info_cache' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.320061] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb8e5c7-a8fa-4207-9f3c-cf9d3583ba66 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.356078] env[62368]: DEBUG oslo_vmware.api [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198502, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.356779] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198505, 'name': Rename_Task, 'duration_secs': 0.194583} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.357155] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 827.357489] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63ed281d-5cdd-4a68-8490-669f075c710d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.369329] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 827.369329] env[62368]: value = "task-1198506" [ 827.369329] env[62368]: _type = "Task" [ 827.369329] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.381468] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198506, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.401438] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cfb97c-28ae-f2b9-1872-12d7edbaff07, 'name': SearchDatastore_Task, 'duration_secs': 0.020297} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.402086] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.402262] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] cda0cf42-11ae-4da3-b838-5bac4bc2f4d1/cda0cf42-11ae-4da3-b838-5bac4bc2f4d1.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.402596] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c36524db-1f76-469f-9e23-b3283f342b1a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.412656] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 827.412656] env[62368]: value = "task-1198507" [ 827.412656] env[62368]: _type = "Task" [ 827.412656] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.427300] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198507, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.660573] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.730843] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.731624] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.731624] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.732975] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.732975] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.732975] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.732975] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.732975] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.732975] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.733219] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.734114] env[62368]: DEBUG nova.virt.hardware [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.734624] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ff25cf-8fcc-47d9-80b7-ee6b5659c8f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.745619] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230663a6-0097-4826-bcee-48f1a61a3c54 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.794680] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.163s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.798376] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.840s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.798376] env[62368]: DEBUG nova.objects.instance [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lazy-loading 'resources' on Instance uuid 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.834181] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 827.834645] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-28a5a613-1681-4a14-8404-9aae5938f7d1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.849576] env[62368]: INFO nova.scheduler.client.report [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted allocations for instance 3f1af54b-392f-432a-9ffa-a133da428f94 [ 827.862991] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 827.862991] env[62368]: value = "task-1198508" [ 827.862991] env[62368]: _type = "Task" [ 827.862991] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.863301] env[62368]: DEBUG oslo_vmware.api [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198502, 'name': ReconfigVM_Task, 'duration_secs': 1.098242} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.864532] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.864532] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Reconfigured VM to attach interface {{(pid=62368) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 827.888801] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198506, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.890312] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198508, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.926509] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198507, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.062044] env[62368]: DEBUG nova.compute.manager [req-3728e5d1-9d17-485c-9c80-5c949db3e0b9 req-fcfb81ba-f409-4ad2-a88e-7c13ca9cafe2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-changed-90ff566e-24b4-4a83-a965-0bc25b1152e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.062480] env[62368]: DEBUG nova.compute.manager [req-3728e5d1-9d17-485c-9c80-5c949db3e0b9 req-fcfb81ba-f409-4ad2-a88e-7c13ca9cafe2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Refreshing instance network info cache due to event network-changed-90ff566e-24b4-4a83-a965-0bc25b1152e0. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 828.062779] env[62368]: DEBUG oslo_concurrency.lockutils [req-3728e5d1-9d17-485c-9c80-5c949db3e0b9 req-fcfb81ba-f409-4ad2-a88e-7c13ca9cafe2 service nova] Acquiring lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.062941] env[62368]: DEBUG oslo_concurrency.lockutils [req-3728e5d1-9d17-485c-9c80-5c949db3e0b9 req-fcfb81ba-f409-4ad2-a88e-7c13ca9cafe2 service nova] Acquired lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.063123] env[62368]: DEBUG nova.network.neutron [req-3728e5d1-9d17-485c-9c80-5c949db3e0b9 req-fcfb81ba-f409-4ad2-a88e-7c13ca9cafe2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Refreshing network info cache for port 90ff566e-24b4-4a83-a965-0bc25b1152e0 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 828.341861] env[62368]: DEBUG nova.network.neutron [-] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.375795] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0e0c97dd-f586-4771-9eb6-75fe39e85cd7 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.251s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.380040] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b1e561fe-471e-4811-8170-b7962ef140a4 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "3f1af54b-392f-432a-9ffa-a133da428f94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.804s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.390933] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198508, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.398724] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198506, 'name': PowerOnVM_Task, 'duration_secs': 0.649466} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.399265] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 828.399445] env[62368]: INFO nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Took 8.94 seconds to spawn the instance on the hypervisor. [ 828.399660] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.400708] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54621937-46ab-4c70-bb1e-9a10391d061e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.426276] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198507, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597418} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.426667] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] cda0cf42-11ae-4da3-b838-5bac4bc2f4d1/cda0cf42-11ae-4da3-b838-5bac4bc2f4d1.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.426889] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.427165] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e668897-b893-45f8-bb01-8f1b70dca3da {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.438562] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 828.438562] env[62368]: value = "task-1198509" [ 828.438562] env[62368]: _type = "Task" [ 828.438562] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.449074] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.457062] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Successfully updated port: b65f0e56-58ab-48a2-90ca-ca285f331170 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.692520] env[62368]: DEBUG nova.compute.manager [req-1a70caa7-5069-4c7d-9dd4-5ab2709a838a req-555c29e3-e461-46da-af27-6545eaaef605 service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Received event network-vif-plugged-b65f0e56-58ab-48a2-90ca-ca285f331170 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.692698] env[62368]: DEBUG oslo_concurrency.lockutils [req-1a70caa7-5069-4c7d-9dd4-5ab2709a838a req-555c29e3-e461-46da-af27-6545eaaef605 service nova] Acquiring lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.692914] env[62368]: DEBUG oslo_concurrency.lockutils [req-1a70caa7-5069-4c7d-9dd4-5ab2709a838a req-555c29e3-e461-46da-af27-6545eaaef605 service nova] Lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.693269] env[62368]: DEBUG oslo_concurrency.lockutils [req-1a70caa7-5069-4c7d-9dd4-5ab2709a838a req-555c29e3-e461-46da-af27-6545eaaef605 service nova] Lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.693340] env[62368]: DEBUG nova.compute.manager [req-1a70caa7-5069-4c7d-9dd4-5ab2709a838a req-555c29e3-e461-46da-af27-6545eaaef605 service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] No waiting events found dispatching network-vif-plugged-b65f0e56-58ab-48a2-90ca-ca285f331170 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.693563] env[62368]: WARNING nova.compute.manager [req-1a70caa7-5069-4c7d-9dd4-5ab2709a838a req-555c29e3-e461-46da-af27-6545eaaef605 service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Received unexpected event network-vif-plugged-b65f0e56-58ab-48a2-90ca-ca285f331170 for instance with vm_state building and task_state spawning. [ 828.763428] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381a84eb-716d-454b-b90e-f256e82819e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.773175] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc30954-024f-493e-b4ac-6e7bb6bdd9ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.809760] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697f0e2a-d70c-4e6d-bf64-0b474e5f4ae3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.819387] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb5d99d-bfc3-459c-a457-e38568af2114 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.835906] env[62368]: DEBUG nova.compute.provider_tree [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.845552] env[62368]: INFO nova.compute.manager [-] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Took 1.88 seconds to deallocate network for instance. [ 828.882048] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198508, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.928600] env[62368]: INFO nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Took 38.53 seconds to build instance. [ 828.950941] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197852} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.951181] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.952295] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33432832-cd1e-4c79-aa29-14474e54afc8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.959105] env[62368]: DEBUG nova.network.neutron [req-3728e5d1-9d17-485c-9c80-5c949db3e0b9 req-fcfb81ba-f409-4ad2-a88e-7c13ca9cafe2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updated VIF entry in instance network info cache for port 90ff566e-24b4-4a83-a965-0bc25b1152e0. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 828.959105] env[62368]: DEBUG nova.network.neutron [req-3728e5d1-9d17-485c-9c80-5c949db3e0b9 req-fcfb81ba-f409-4ad2-a88e-7c13ca9cafe2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90ff566e-24b4-4a83-a965-0bc25b1152e0", "address": "fa:16:3e:29:10:46", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ff566e-24", "ovs_interfaceid": "90ff566e-24b4-4a83-a965-0bc25b1152e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.960592] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "refresh_cache-e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.960728] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "refresh_cache-e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.960868] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.991583] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] cda0cf42-11ae-4da3-b838-5bac4bc2f4d1/cda0cf42-11ae-4da3-b838-5bac4bc2f4d1.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.992344] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f22ce425-db78-4287-b1e4-5f511dd82646 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.021584] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 829.021584] env[62368]: value = "task-1198510" [ 829.021584] env[62368]: _type = "Task" [ 829.021584] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.032691] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198510, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.197517] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.342227] env[62368]: DEBUG nova.scheduler.client.report [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.355733] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.384465] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198508, 'name': CreateSnapshot_Task, 'duration_secs': 1.35551} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.384465] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 829.384930] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf3cc56-9574-4ca7-94f1-b947255b0478 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.432697] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.044s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.470101] env[62368]: DEBUG oslo_concurrency.lockutils [req-3728e5d1-9d17-485c-9c80-5c949db3e0b9 req-fcfb81ba-f409-4ad2-a88e-7c13ca9cafe2 service nova] Releasing lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.533888] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198510, 'name': ReconfigVM_Task, 'duration_secs': 0.511648} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.534281] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Reconfigured VM instance instance-00000045 to attach disk [datastore2] cda0cf42-11ae-4da3-b838-5bac4bc2f4d1/cda0cf42-11ae-4da3-b838-5bac4bc2f4d1.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.535019] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-525e70c7-5141-4cdc-bf7a-050983c61974 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.540936] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.544706] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 829.544706] env[62368]: value = "task-1198511" [ 829.544706] env[62368]: _type = "Task" [ 829.544706] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.555089] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198511, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.699359] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Releasing lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.699613] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updated the network info_cache for instance {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 829.699797] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.700210] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.700210] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.700420] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.700420] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.700693] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.700693] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 829.700948] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.847490] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.050s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.851167] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.839s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.855189] env[62368]: INFO nova.compute.claims [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.884876] env[62368]: INFO nova.scheduler.client.report [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Deleted allocations for instance 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d [ 829.894979] env[62368]: DEBUG nova.network.neutron [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Updating instance_info_cache with network_info: [{"id": "b65f0e56-58ab-48a2-90ca-ca285f331170", "address": "fa:16:3e:af:4e:47", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb65f0e56-58", "ovs_interfaceid": "b65f0e56-58ab-48a2-90ca-ca285f331170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.913234] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 829.915010] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-96a463d6-a6b1-4a52-9f8d-2ca7a9159c83 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.930183] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 829.930183] env[62368]: value = "task-1198512" [ 829.930183] env[62368]: _type = "Task" [ 829.930183] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.954084] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198512, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.058640] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198511, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.209543] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.244088] env[62368]: DEBUG nova.compute.manager [req-631f97c3-2063-4339-8dff-ba3a2dfd1259 req-881a905d-6187-462b-a3cc-6ff913111e0e service nova] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Received event network-vif-deleted-3ba9f19e-f789-4e0a-b733-bba38fc85baf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.398587] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cbc0a815-2bde-4926-a8bc-a15fc2a45984 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.626s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.399286] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "refresh_cache-e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.399571] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Instance network_info: |[{"id": "b65f0e56-58ab-48a2-90ca-ca285f331170", "address": "fa:16:3e:af:4e:47", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb65f0e56-58", "ovs_interfaceid": "b65f0e56-58ab-48a2-90ca-ca285f331170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.400157] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:4e:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13b62154-a0e1-4eed-bc30-6464b15993bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b65f0e56-58ab-48a2-90ca-ca285f331170', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.408409] env[62368]: DEBUG oslo.service.loopingcall [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.408648] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.408897] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7ca0449-8e6f-458a-9821-c4c082d4f7c1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.431356] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.431356] env[62368]: value = "task-1198513" [ 830.431356] env[62368]: _type = "Task" [ 830.431356] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.446548] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198512, 'name': CloneVM_Task} progress is 93%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.450148] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198513, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.559694] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198511, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.626097] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-c582c777-e544-40af-9f04-d8c171475bc1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.627233] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-c582c777-e544-40af-9f04-d8c171475bc1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.627233] env[62368]: DEBUG nova.objects.instance [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'flavor' on Instance uuid 67d8c43a-d3af-407b-847a-1c2c8c53dbfb {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.708441] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "b87094da-6258-469e-ab37-5557955ad3a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.708777] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "b87094da-6258-469e-ab37-5557955ad3a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.953927] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198513, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.960257] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198512, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.061052] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198511, 'name': Rename_Task, 'duration_secs': 1.363962} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.061280] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 831.061569] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a81d6a14-d6b3-4efd-8ffe-0309542f4fa5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.070635] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 831.070635] env[62368]: value = "task-1198514" [ 831.070635] env[62368]: _type = "Task" [ 831.070635] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.082799] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198514, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.087833] env[62368]: DEBUG nova.compute.manager [req-6492a807-b8e8-45be-b77a-777af1759bea req-9f7f31b4-0553-4d34-862e-93587852e75c service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Received event network-changed-b65f0e56-58ab-48a2-90ca-ca285f331170 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.088045] env[62368]: DEBUG nova.compute.manager [req-6492a807-b8e8-45be-b77a-777af1759bea req-9f7f31b4-0553-4d34-862e-93587852e75c service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Refreshing instance network info cache due to event network-changed-b65f0e56-58ab-48a2-90ca-ca285f331170. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.088284] env[62368]: DEBUG oslo_concurrency.lockutils [req-6492a807-b8e8-45be-b77a-777af1759bea req-9f7f31b4-0553-4d34-862e-93587852e75c service nova] Acquiring lock "refresh_cache-e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.088430] env[62368]: DEBUG oslo_concurrency.lockutils [req-6492a807-b8e8-45be-b77a-777af1759bea req-9f7f31b4-0553-4d34-862e-93587852e75c service nova] Acquired lock "refresh_cache-e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.088593] env[62368]: DEBUG nova.network.neutron [req-6492a807-b8e8-45be-b77a-777af1759bea req-9f7f31b4-0553-4d34-862e-93587852e75c service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Refreshing network info cache for port b65f0e56-58ab-48a2-90ca-ca285f331170 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 831.211586] env[62368]: DEBUG nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.273229] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4781c3f-599d-4111-9923-6f52bcc37c44 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.285428] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e5f3ef-e68f-46e1-8d4a-f9020473d21f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.319498] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4529fd-1b41-456c-b44d-6422ac0f1069 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.328360] env[62368]: DEBUG nova.objects.instance [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'pci_requests' on Instance uuid 67d8c43a-d3af-407b-847a-1c2c8c53dbfb {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.330448] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be074f7b-d824-4045-9d3e-1a53b0759fb2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.347599] env[62368]: DEBUG nova.compute.provider_tree [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.392395] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "69ddb565-6c79-44e5-a7d1-d339ab426fae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.392807] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "69ddb565-6c79-44e5-a7d1-d339ab426fae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.445594] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198512, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.456398] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198513, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.582578] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198514, 'name': PowerOnVM_Task} progress is 86%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.741999] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.835971] env[62368]: DEBUG nova.objects.base [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Object Instance<67d8c43a-d3af-407b-847a-1c2c8c53dbfb> lazy-loaded attributes: flavor,pci_requests {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 831.836314] env[62368]: DEBUG nova.network.neutron [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 831.856029] env[62368]: DEBUG nova.scheduler.client.report [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.925103] env[62368]: DEBUG nova.network.neutron [req-6492a807-b8e8-45be-b77a-777af1759bea req-9f7f31b4-0553-4d34-862e-93587852e75c service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Updated VIF entry in instance network info cache for port b65f0e56-58ab-48a2-90ca-ca285f331170. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.925647] env[62368]: DEBUG nova.network.neutron [req-6492a807-b8e8-45be-b77a-777af1759bea req-9f7f31b4-0553-4d34-862e-93587852e75c service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Updating instance_info_cache with network_info: [{"id": "b65f0e56-58ab-48a2-90ca-ca285f331170", "address": "fa:16:3e:af:4e:47", "network": {"id": "0c508a5e-fe3a-4477-a1f8-c57e37187b22", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1097319401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bc06cf02fa4cebb141a45d732b123a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb65f0e56-58", "ovs_interfaceid": "b65f0e56-58ab-48a2-90ca-ca285f331170", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.946412] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198512, 'name': CloneVM_Task, 'duration_secs': 1.998359} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.949456] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Created linked-clone VM from snapshot [ 831.951355] env[62368]: DEBUG nova.policy [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68c2209271e043b698fd3794955f40bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcecae6c09124e9fa4dcd5d63833b0bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.954270] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019bb4e1-4b43-4c47-9516-ccf1da4ae523 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.967386] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198513, 'name': CreateVM_Task, 'duration_secs': 1.513839} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.968149] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Uploading image 910933e8-6f2d-4e86-b7c1-efe87e2bffff {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 831.971953] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 831.971953] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.971953] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.971953] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.971953] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2405915b-5091-41ef-8d61-963b4235fc11 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.980343] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 831.980343] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522c3feb-0473-bb5b-ad58-03ebe772a7cf" [ 831.980343] env[62368]: _type = "Task" [ 831.980343] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.986120] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 831.986538] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1d236c82-327f-4ca6-a99d-c90ea82dcc8a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.992275] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522c3feb-0473-bb5b-ad58-03ebe772a7cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.999393] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 831.999393] env[62368]: value = "task-1198515" [ 831.999393] env[62368]: _type = "Task" [ 831.999393] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.010580] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198515, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.081201] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198514, 'name': PowerOnVM_Task, 'duration_secs': 0.834228} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.081898] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 832.082222] env[62368]: INFO nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Took 10.02 seconds to spawn the instance on the hypervisor. [ 832.082324] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.083193] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723c53b4-064f-4258-a40c-5e9947ebee6c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.370340] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.370888] env[62368]: DEBUG nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.373622] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.730s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.373851] env[62368]: DEBUG nova.objects.instance [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lazy-loading 'resources' on Instance uuid 87dc392f-7bba-44bb-ba38-2fd4e1c6129e {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.429064] env[62368]: DEBUG oslo_concurrency.lockutils [req-6492a807-b8e8-45be-b77a-777af1759bea req-9f7f31b4-0553-4d34-862e-93587852e75c service nova] Releasing lock "refresh_cache-e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.490299] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522c3feb-0473-bb5b-ad58-03ebe772a7cf, 'name': SearchDatastore_Task, 'duration_secs': 0.014879} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.490514] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.490741] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.490980] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.491149] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.491332] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.491603] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afe6be15-2d7f-4373-b12a-7468f2e98ffb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.504222] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.504430] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.505527] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1360be8-d580-4c5f-9d0a-550d78589f82 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.513089] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198515, 'name': Destroy_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.517254] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 832.517254] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5210c27a-d2b9-a897-fe4b-468a38c98fc1" [ 832.517254] env[62368]: _type = "Task" [ 832.517254] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.526031] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5210c27a-d2b9-a897-fe4b-468a38c98fc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.609266] env[62368]: INFO nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Took 40.46 seconds to build instance. [ 832.880333] env[62368]: DEBUG nova.compute.utils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.881974] env[62368]: DEBUG nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.882176] env[62368]: DEBUG nova.network.neutron [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 832.922858] env[62368]: DEBUG nova.policy [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87c366f937e2485c8b0dc317d6acb8f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '430e72f3b08a4207a2fb4465ff160714', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.015143] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198515, 'name': Destroy_Task, 'duration_secs': 0.556679} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.015439] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Destroyed the VM [ 833.015708] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 833.015975] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-34b7fc1a-2498-4de1-b7af-3ce66b36e637 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.027522] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 833.027522] env[62368]: value = "task-1198516" [ 833.027522] env[62368]: _type = "Task" [ 833.027522] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.031628] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5210c27a-d2b9-a897-fe4b-468a38c98fc1, 'name': SearchDatastore_Task, 'duration_secs': 0.017039} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.038408] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d95a0a9b-0648-4a02-aa15-141a34313486 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.048898] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198516, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.049344] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 833.049344] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f93ebc-1cf5-ed82-934d-4224a78088ec" [ 833.049344] env[62368]: _type = "Task" [ 833.049344] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.062407] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f93ebc-1cf5-ed82-934d-4224a78088ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.112115] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.688s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.242029] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cd7821-eefd-4b6f-a628-bb19afa4443d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.252428] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba5d39a-0aa3-4a8e-a465-f65db10e408f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.294160] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355ef155-96b2-40e7-a409-20c6b8c31265 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.307031] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88af7fa5-a74e-44f2-8368-8f969e867d6e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.321854] env[62368]: DEBUG nova.compute.provider_tree [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.359409] env[62368]: DEBUG nova.network.neutron [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Successfully created port: b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.384997] env[62368]: DEBUG nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.542266] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198516, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.566066] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f93ebc-1cf5-ed82-934d-4224a78088ec, 'name': SearchDatastore_Task, 'duration_secs': 0.015896} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.566360] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.566659] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f/e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 833.566938] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-484f842a-cc1f-46b7-9109-2190aaea5f6b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.580986] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 833.580986] env[62368]: value = "task-1198517" [ 833.580986] env[62368]: _type = "Task" [ 833.580986] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.587123] env[62368]: DEBUG nova.network.neutron [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Successfully updated port: c582c777-e544-40af-9f04-d8c171475bc1 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.594302] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198517, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.618599] env[62368]: DEBUG nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.667199] env[62368]: DEBUG nova.compute.manager [req-dc1e59ea-5cb9-4243-a6a0-55b6afcee1fd req-33116fe3-ad24-496c-aa9f-6dced35bd3f3 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-vif-plugged-c582c777-e544-40af-9f04-d8c171475bc1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.667199] env[62368]: DEBUG oslo_concurrency.lockutils [req-dc1e59ea-5cb9-4243-a6a0-55b6afcee1fd req-33116fe3-ad24-496c-aa9f-6dced35bd3f3 service nova] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.667199] env[62368]: DEBUG oslo_concurrency.lockutils [req-dc1e59ea-5cb9-4243-a6a0-55b6afcee1fd req-33116fe3-ad24-496c-aa9f-6dced35bd3f3 service nova] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.667199] env[62368]: DEBUG oslo_concurrency.lockutils [req-dc1e59ea-5cb9-4243-a6a0-55b6afcee1fd req-33116fe3-ad24-496c-aa9f-6dced35bd3f3 service nova] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.667199] env[62368]: DEBUG nova.compute.manager [req-dc1e59ea-5cb9-4243-a6a0-55b6afcee1fd req-33116fe3-ad24-496c-aa9f-6dced35bd3f3 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] No waiting events found dispatching network-vif-plugged-c582c777-e544-40af-9f04-d8c171475bc1 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.667199] env[62368]: WARNING nova.compute.manager [req-dc1e59ea-5cb9-4243-a6a0-55b6afcee1fd req-33116fe3-ad24-496c-aa9f-6dced35bd3f3 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received unexpected event network-vif-plugged-c582c777-e544-40af-9f04-d8c171475bc1 for instance with vm_state active and task_state None. [ 833.827648] env[62368]: DEBUG nova.scheduler.client.report [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.046277] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198516, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.093318] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198517, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.096463] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.097078] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.097189] env[62368]: DEBUG nova.network.neutron [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 834.143643] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.333422] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.337783] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 31.333s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.354842] env[62368]: INFO nova.scheduler.client.report [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Deleted allocations for instance 87dc392f-7bba-44bb-ba38-2fd4e1c6129e [ 834.366541] env[62368]: DEBUG oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d7488d-1323-8742-5da7-9e3c7fdeea0c/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 834.367597] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58db0c43-e7e7-4fba-a167-b3a55bf86bc3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.376438] env[62368]: DEBUG oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d7488d-1323-8742-5da7-9e3c7fdeea0c/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 834.376689] env[62368]: ERROR oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d7488d-1323-8742-5da7-9e3c7fdeea0c/disk-0.vmdk due to incomplete transfer. [ 834.376933] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-536feaf1-cf19-43fa-a899-919e4941481c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.384706] env[62368]: DEBUG oslo_vmware.rw_handles [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d7488d-1323-8742-5da7-9e3c7fdeea0c/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 834.384937] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Uploaded image ba6b39d1-f7e6-4aec-93f4-f648afffeff8 to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 834.388142] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 834.388446] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0ce01713-2bee-4b7d-b36a-867abffd3194 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.396035] env[62368]: DEBUG nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.399041] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 834.399041] env[62368]: value = "task-1198518" [ 834.399041] env[62368]: _type = "Task" [ 834.399041] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.409261] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198518, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.421217] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.421510] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.421700] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.421899] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.422062] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.422221] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.422513] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.422603] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.422775] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.422961] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.423142] env[62368]: DEBUG nova.virt.hardware [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.424025] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07594174-81ef-41ee-9093-6589f32aa431 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.431614] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0401f84a-9268-46b2-8bcf-309c32872345 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.543180] env[62368]: DEBUG oslo_vmware.api [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198516, 'name': RemoveSnapshot_Task, 'duration_secs': 1.076661} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.543526] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 834.593114] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198517, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652993} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.593388] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f/e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.593606] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.593864] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9446c749-665f-4c5d-92df-3d398cdd0304 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.600751] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 834.600751] env[62368]: value = "task-1198519" [ 834.600751] env[62368]: _type = "Task" [ 834.600751] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.610491] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198519, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.642598] env[62368]: WARNING nova.network.neutron [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] 3ee164de-30af-4404-935a-a7f988d824b1 already exists in list: networks containing: ['3ee164de-30af-4404-935a-a7f988d824b1']. ignoring it [ 834.642815] env[62368]: WARNING nova.network.neutron [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] 3ee164de-30af-4404-935a-a7f988d824b1 already exists in list: networks containing: ['3ee164de-30af-4404-935a-a7f988d824b1']. ignoring it [ 834.845851] env[62368]: INFO nova.compute.claims [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.867389] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce90fc72-8946-4028-96d9-85e7dad88302 tempest-MultipleCreateTestJSON-1987620845 tempest-MultipleCreateTestJSON-1987620845-project-member] Lock "87dc392f-7bba-44bb-ba38-2fd4e1c6129e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.637s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.909834] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198518, 'name': Destroy_Task, 'duration_secs': 0.362932} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.911100] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Destroyed the VM [ 834.911100] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 834.911871] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1ea512ae-1f69-4234-8fd7-7cb09cfc9780 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.918089] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 834.918089] env[62368]: value = "task-1198520" [ 834.918089] env[62368]: _type = "Task" [ 834.918089] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.927894] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198520, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.045682] env[62368]: DEBUG nova.compute.manager [req-470a8b19-1bd9-4110-96cf-70d471ac6b78 req-1485faae-a11f-468d-9cf6-9fb0f887e5c9 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Received event network-vif-plugged-b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.045908] env[62368]: DEBUG oslo_concurrency.lockutils [req-470a8b19-1bd9-4110-96cf-70d471ac6b78 req-1485faae-a11f-468d-9cf6-9fb0f887e5c9 service nova] Acquiring lock "2a46c954-449e-4d62-be80-add1040ed4c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.046188] env[62368]: DEBUG oslo_concurrency.lockutils [req-470a8b19-1bd9-4110-96cf-70d471ac6b78 req-1485faae-a11f-468d-9cf6-9fb0f887e5c9 service nova] Lock "2a46c954-449e-4d62-be80-add1040ed4c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.046400] env[62368]: DEBUG oslo_concurrency.lockutils [req-470a8b19-1bd9-4110-96cf-70d471ac6b78 req-1485faae-a11f-468d-9cf6-9fb0f887e5c9 service nova] Lock "2a46c954-449e-4d62-be80-add1040ed4c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.048067] env[62368]: DEBUG nova.compute.manager [req-470a8b19-1bd9-4110-96cf-70d471ac6b78 req-1485faae-a11f-468d-9cf6-9fb0f887e5c9 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] No waiting events found dispatching network-vif-plugged-b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.048067] env[62368]: WARNING nova.compute.manager [req-470a8b19-1bd9-4110-96cf-70d471ac6b78 req-1485faae-a11f-468d-9cf6-9fb0f887e5c9 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Received unexpected event network-vif-plugged-b4fab9b2-299a-4e59-92bf-efdc3d583416 for instance with vm_state building and task_state spawning. [ 835.048347] env[62368]: WARNING nova.compute.manager [None req-6ccabf5b-e6f9-46da-9d4d-83665270fbf3 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Image not found during snapshot: nova.exception.ImageNotFound: Image 910933e8-6f2d-4e86-b7c1-efe87e2bffff could not be found. [ 835.079143] env[62368]: DEBUG nova.network.neutron [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Successfully updated port: b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.112327] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198519, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.186301} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.112941] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.113803] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e3a3f4-706c-45ce-aec7-5fe4db482af8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.140845] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f/e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.142197] env[62368]: DEBUG nova.network.neutron [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90ff566e-24b4-4a83-a965-0bc25b1152e0", "address": "fa:16:3e:29:10:46", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ff566e-24", "ovs_interfaceid": "90ff566e-24b4-4a83-a965-0bc25b1152e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c582c777-e544-40af-9f04-d8c171475bc1", "address": "fa:16:3e:e4:28:b5", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc582c777-e5", "ovs_interfaceid": "c582c777-e544-40af-9f04-d8c171475bc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.143496] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-195cd269-0c57-44d7-86ae-e509a2965b22 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.159374] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.160491] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.160491] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.161304] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b809d37-fcd2-46f6-80ad-e08c9e903e30 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.184329] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 835.184329] env[62368]: value = "task-1198521" [ 835.184329] env[62368]: _type = "Task" [ 835.184329] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.184844] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.185183] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.185451] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.185676] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.185828] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.185977] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.186200] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.186360] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.186525] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.186687] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.186984] env[62368]: DEBUG nova.virt.hardware [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.193332] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Reconfiguring VM to attach interface {{(pid=62368) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 835.193899] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6686a97f-e3f4-41a3-aab9-9ecb12726ab4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.214480] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198521, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.215674] env[62368]: DEBUG oslo_vmware.api [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 835.215674] env[62368]: value = "task-1198522" [ 835.215674] env[62368]: _type = "Task" [ 835.215674] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.223356] env[62368]: DEBUG oslo_vmware.api [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198522, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.352452] env[62368]: INFO nova.compute.resource_tracker [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating resource usage from migration 97c6a2e1-faef-43ad-8d88-091a4fcca87e [ 835.429540] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198520, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.581195] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.581347] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.581499] env[62368]: DEBUG nova.network.neutron [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 835.690364] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03c2f72-f47c-4dc7-b79d-e34a7e0a7485 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.700816] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60f07fd-1790-4f7a-98ac-19379c6e71fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.707398] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198521, 'name': ReconfigVM_Task, 'duration_secs': 0.356691} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.708179] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Reconfigured VM instance instance-00000046 to attach disk [datastore2] e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f/e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.708898] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83cc18a8-0064-48a9-a4a7-d45dd50d395a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.752692] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6858f4de-4333-4dd8-928b-09caa2976264 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.758124] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 835.758124] env[62368]: value = "task-1198523" [ 835.758124] env[62368]: _type = "Task" [ 835.758124] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.771257] env[62368]: DEBUG oslo_vmware.api [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198522, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.773910] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a30127a-e56d-484c-a1c9-e35d2c2321e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.785787] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198523, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.801472] env[62368]: DEBUG nova.compute.provider_tree [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.928961] env[62368]: DEBUG oslo_vmware.api [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198520, 'name': RemoveSnapshot_Task, 'duration_secs': 0.53291} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.929253] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 835.929495] env[62368]: INFO nova.compute.manager [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Took 15.78 seconds to snapshot the instance on the hypervisor. [ 835.962823] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "d865d433-2341-4f6c-b840-609530ae2e51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.963094] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "d865d433-2341-4f6c-b840-609530ae2e51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.963321] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "d865d433-2341-4f6c-b840-609530ae2e51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.963501] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "d865d433-2341-4f6c-b840-609530ae2e51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.963678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "d865d433-2341-4f6c-b840-609530ae2e51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.965806] env[62368]: INFO nova.compute.manager [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Terminating instance [ 835.967709] env[62368]: DEBUG nova.compute.manager [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 835.967912] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 835.968977] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db53bb8-5d55-41eb-87bf-ef30b3c58758 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.976475] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 835.976719] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e6eef94-79d8-4137-8876-414cb80cc2d1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.983281] env[62368]: DEBUG oslo_vmware.api [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 835.983281] env[62368]: value = "task-1198524" [ 835.983281] env[62368]: _type = "Task" [ 835.983281] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.991977] env[62368]: DEBUG oslo_vmware.api [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.118713] env[62368]: DEBUG nova.network.neutron [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.256363] env[62368]: DEBUG oslo_vmware.api [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198522, 'name': ReconfigVM_Task, 'duration_secs': 0.649081} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.259224] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.259509] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Reconfigured VM to attach interface {{(pid=62368) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 836.272907] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198523, 'name': Rename_Task, 'duration_secs': 0.23372} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.273258] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.273548] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1248bcde-4c5d-4283-97f4-35511f6f5744 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.281309] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 836.281309] env[62368]: value = "task-1198525" [ 836.281309] env[62368]: _type = "Task" [ 836.281309] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.293783] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198525, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.302728] env[62368]: DEBUG nova.network.neutron [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updating instance_info_cache with network_info: [{"id": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "address": "fa:16:3e:2d:06:3b", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4fab9b2-29", "ovs_interfaceid": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.304214] env[62368]: DEBUG nova.scheduler.client.report [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.377121] env[62368]: DEBUG nova.compute.manager [req-2b2656ea-a74f-4098-999e-b5c5bd3a5a6b req-ed9c0e55-17db-4c64-9a53-0640814d8180 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-changed-c582c777-e544-40af-9f04-d8c171475bc1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.377608] env[62368]: DEBUG nova.compute.manager [req-2b2656ea-a74f-4098-999e-b5c5bd3a5a6b req-ed9c0e55-17db-4c64-9a53-0640814d8180 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Refreshing instance network info cache due to event network-changed-c582c777-e544-40af-9f04-d8c171475bc1. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.377996] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b2656ea-a74f-4098-999e-b5c5bd3a5a6b req-ed9c0e55-17db-4c64-9a53-0640814d8180 service nova] Acquiring lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.378302] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b2656ea-a74f-4098-999e-b5c5bd3a5a6b req-ed9c0e55-17db-4c64-9a53-0640814d8180 service nova] Acquired lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.378605] env[62368]: DEBUG nova.network.neutron [req-2b2656ea-a74f-4098-999e-b5c5bd3a5a6b req-ed9c0e55-17db-4c64-9a53-0640814d8180 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Refreshing network info cache for port c582c777-e544-40af-9f04-d8c171475bc1 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 836.473031] env[62368]: DEBUG nova.compute.manager [None req-244ef2cc-693a-4f52-af5e-486e47bc057f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Found 2 images (rotation: 2) {{(pid=62368) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 836.492431] env[62368]: DEBUG oslo_vmware.api [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198524, 'name': PowerOffVM_Task, 'duration_secs': 0.255519} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.492937] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 836.493245] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 836.493627] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6302ffd-8e2f-494c-b7f2-79f8becd66ef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.653020] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 836.653020] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 836.653020] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleting the datastore file [datastore1] d865d433-2341-4f6c-b840-609530ae2e51 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 836.653020] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b08f3c6-784d-4cdc-ba65-0b0063a0a52c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.659177] env[62368]: DEBUG oslo_vmware.api [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 836.659177] env[62368]: value = "task-1198527" [ 836.659177] env[62368]: _type = "Task" [ 836.659177] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.667272] env[62368]: DEBUG oslo_vmware.api [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198527, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.769019] env[62368]: DEBUG oslo_concurrency.lockutils [None req-696068c3-03c4-48e5-a666-ff5200c4c114 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-c582c777-e544-40af-9f04-d8c171475bc1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.138s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.793625] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198525, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.813264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.813264] env[62368]: DEBUG nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Instance network_info: |[{"id": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "address": "fa:16:3e:2d:06:3b", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4fab9b2-29", "ovs_interfaceid": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.813264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.474s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.813264] env[62368]: INFO nova.compute.manager [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Migrating [ 836.813264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.813264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.814175] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:06:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4fab9b2-299a-4e59-92bf-efdc3d583416', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.825321] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Creating folder: Project (430e72f3b08a4207a2fb4465ff160714). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 836.825832] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.452s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.827593] env[62368]: INFO nova.compute.claims [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.831905] env[62368]: INFO nova.compute.rpcapi [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 836.832447] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.842487] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8bb2bb9b-e7d1-4d6c-8657-8ae91ca0e6b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.863621] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Created folder: Project (430e72f3b08a4207a2fb4465ff160714) in parent group-v259706. [ 836.863621] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Creating folder: Instances. Parent ref: group-v259815. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 836.863621] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ccbeb6e0-3062-4edb-8f24-a39ea82e3b3f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.873464] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Created folder: Instances in parent group-v259815. [ 836.873716] env[62368]: DEBUG oslo.service.loopingcall [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.873987] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 836.874395] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-063e7667-4059-4383-a137-4a96b42c301a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.896539] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.896539] env[62368]: value = "task-1198530" [ 836.896539] env[62368]: _type = "Task" [ 836.896539] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.904957] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198530, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.148451] env[62368]: DEBUG nova.network.neutron [req-2b2656ea-a74f-4098-999e-b5c5bd3a5a6b req-ed9c0e55-17db-4c64-9a53-0640814d8180 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updated VIF entry in instance network info cache for port c582c777-e544-40af-9f04-d8c171475bc1. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 837.149092] env[62368]: DEBUG nova.network.neutron [req-2b2656ea-a74f-4098-999e-b5c5bd3a5a6b req-ed9c0e55-17db-4c64-9a53-0640814d8180 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "90ff566e-24b4-4a83-a965-0bc25b1152e0", "address": "fa:16:3e:29:10:46", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ff566e-24", "ovs_interfaceid": "90ff566e-24b4-4a83-a965-0bc25b1152e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c582c777-e544-40af-9f04-d8c171475bc1", "address": "fa:16:3e:e4:28:b5", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc582c777-e5", "ovs_interfaceid": "c582c777-e544-40af-9f04-d8c171475bc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.173516] env[62368]: DEBUG oslo_vmware.api [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198527, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.465595} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.173516] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 837.173516] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 837.173770] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 837.173816] env[62368]: INFO nova.compute.manager [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Took 1.21 seconds to destroy the instance on the hypervisor. [ 837.174095] env[62368]: DEBUG oslo.service.loopingcall [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.174290] env[62368]: DEBUG nova.compute.manager [-] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 837.174393] env[62368]: DEBUG nova.network.neutron [-] [instance: d865d433-2341-4f6c-b840-609530ae2e51] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 837.293969] env[62368]: DEBUG oslo_vmware.api [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198525, 'name': PowerOnVM_Task, 'duration_secs': 0.782626} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.294365] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.294653] env[62368]: INFO nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Took 9.63 seconds to spawn the instance on the hypervisor. [ 837.295048] env[62368]: DEBUG nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.296210] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc76b64-2be8-4b3f-9ad9-7c4d2a82b377 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.356468] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.356468] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.356468] env[62368]: DEBUG nova.network.neutron [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 837.408832] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198530, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.486903] env[62368]: DEBUG nova.compute.manager [req-d74c71fe-610b-4f5f-a231-4ee8c8c81ed0 req-ab283988-6d65-43ff-bb4b-7687abb67a90 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Received event network-changed-b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.487835] env[62368]: DEBUG nova.compute.manager [req-d74c71fe-610b-4f5f-a231-4ee8c8c81ed0 req-ab283988-6d65-43ff-bb4b-7687abb67a90 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Refreshing instance network info cache due to event network-changed-b4fab9b2-299a-4e59-92bf-efdc3d583416. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.488279] env[62368]: DEBUG oslo_concurrency.lockutils [req-d74c71fe-610b-4f5f-a231-4ee8c8c81ed0 req-ab283988-6d65-43ff-bb4b-7687abb67a90 service nova] Acquiring lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.488570] env[62368]: DEBUG oslo_concurrency.lockutils [req-d74c71fe-610b-4f5f-a231-4ee8c8c81ed0 req-ab283988-6d65-43ff-bb4b-7687abb67a90 service nova] Acquired lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.488872] env[62368]: DEBUG nova.network.neutron [req-d74c71fe-610b-4f5f-a231-4ee8c8c81ed0 req-ab283988-6d65-43ff-bb4b-7687abb67a90 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Refreshing network info cache for port b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.657106] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b2656ea-a74f-4098-999e-b5c5bd3a5a6b req-ed9c0e55-17db-4c64-9a53-0640814d8180 service nova] Releasing lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.818125] env[62368]: INFO nova.compute.manager [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Took 39.87 seconds to build instance. [ 837.913977] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198530, 'name': CreateVM_Task, 'duration_secs': 0.567597} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.914181] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 837.915566] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.915566] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.915706] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.915963] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d7d9f6c-dbc0-4c5b-a488-1b8796b7ba97 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.920839] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 837.920839] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529b6f08-ddf3-9c49-e0d2-26d911232194" [ 837.920839] env[62368]: _type = "Task" [ 837.920839] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.930684] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529b6f08-ddf3-9c49-e0d2-26d911232194, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.954752] env[62368]: DEBUG nova.network.neutron [-] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.186244] env[62368]: DEBUG nova.network.neutron [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance_info_cache with network_info: [{"id": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "address": "fa:16:3e:98:a0:d2", "network": {"id": "34f156be-cab3-45a2-8b52-74dd4cd1e42d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2a054e975862463bb842479ad19393ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap444a47e3-2f", "ovs_interfaceid": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.293639] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0124e3ca-87d3-4f28-9b41-3612087be77e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.304069] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944fe49c-58d9-4751-825c-cb25859d3839 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.337238] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a8812592-1d5c-4058-8320-38b6bff9a634 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.886s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.338239] env[62368]: DEBUG nova.network.neutron [req-d74c71fe-610b-4f5f-a231-4ee8c8c81ed0 req-ab283988-6d65-43ff-bb4b-7687abb67a90 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updated VIF entry in instance network info cache for port b4fab9b2-299a-4e59-92bf-efdc3d583416. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 838.338700] env[62368]: DEBUG nova.network.neutron [req-d74c71fe-610b-4f5f-a231-4ee8c8c81ed0 req-ab283988-6d65-43ff-bb4b-7687abb67a90 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updating instance_info_cache with network_info: [{"id": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "address": "fa:16:3e:2d:06:3b", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4fab9b2-29", "ovs_interfaceid": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.345352] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64d75dc-49c0-4348-82e1-8ed1c26966fb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.353928] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23df791f-2279-4ba8-93cc-6803702e4f71 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.371954] env[62368]: DEBUG nova.compute.provider_tree [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.434412] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529b6f08-ddf3-9c49-e0d2-26d911232194, 'name': SearchDatastore_Task, 'duration_secs': 0.019792} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.434750] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.435048] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.435392] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.435569] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.435870] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.436090] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3bc1d1a9-c19b-4d35-a3e8-0f1248972da7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.447141] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.447350] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 838.448116] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-212da528-bdd9-4a5a-ba8f-f249872923a5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.454875] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 838.454875] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5223e71e-b05c-00b8-852e-049568968ec4" [ 838.454875] env[62368]: _type = "Task" [ 838.454875] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.458133] env[62368]: INFO nova.compute.manager [-] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Took 1.28 seconds to deallocate network for instance. [ 838.466591] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5223e71e-b05c-00b8-852e-049568968ec4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.473447] env[62368]: DEBUG nova.compute.manager [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.474437] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedbd858-b944-420c-9bb3-6c6ccce0f4c3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.589729] env[62368]: DEBUG nova.compute.manager [req-368ad2c6-67e6-4e8d-b42d-3cb45337394b req-3388fea8-e131-4e4c-bc3c-19d023a4e15b service nova] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Received event network-vif-deleted-0f20ad94-5a78-40ab-9ce7-1a56c05d20bc {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.692775] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.848942] env[62368]: DEBUG oslo_concurrency.lockutils [req-d74c71fe-610b-4f5f-a231-4ee8c8c81ed0 req-ab283988-6d65-43ff-bb4b-7687abb67a90 service nova] Releasing lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.878728] env[62368]: DEBUG nova.scheduler.client.report [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.969054] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.969387] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5223e71e-b05c-00b8-852e-049568968ec4, 'name': SearchDatastore_Task, 'duration_secs': 0.020251} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.970200] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d40e20f-0776-4f2c-a434-0610b762e7d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.977513] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 838.977513] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526fd59c-5ed2-9b3e-aa62-f0caa16f2614" [ 838.977513] env[62368]: _type = "Task" [ 838.977513] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.987288] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526fd59c-5ed2-9b3e-aa62-f0caa16f2614, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.988099] env[62368]: INFO nova.compute.manager [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] instance snapshotting [ 838.988682] env[62368]: DEBUG nova.objects.instance [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'flavor' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.301033] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-90ff566e-24b4-4a83-a965-0bc25b1152e0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.301460] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-90ff566e-24b4-4a83-a965-0bc25b1152e0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.383554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.384163] env[62368]: DEBUG nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.386710] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.588s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.388602] env[62368]: INFO nova.compute.claims [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.441408] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.442220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.442220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.442220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.443087] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.445837] env[62368]: INFO nova.compute.manager [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Terminating instance [ 839.448137] env[62368]: DEBUG nova.compute.manager [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.448342] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.449191] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4272627-3ebc-48dc-9dce-5ae1b7d0eb39 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.458094] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.458356] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0aea720-3e6d-402d-887c-df4b9ef04835 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.466045] env[62368]: DEBUG oslo_vmware.api [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 839.466045] env[62368]: value = "task-1198531" [ 839.466045] env[62368]: _type = "Task" [ 839.466045] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.476076] env[62368]: DEBUG oslo_vmware.api [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.490137] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526fd59c-5ed2-9b3e-aa62-f0caa16f2614, 'name': SearchDatastore_Task, 'duration_secs': 0.038078} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.490282] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.490569] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 2a46c954-449e-4d62-be80-add1040ed4c0/2a46c954-449e-4d62-be80-add1040ed4c0.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.490862] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b59a5d75-466d-4e87-a4a8-0b397158593d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.498931] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae9a9f6-ac57-4f72-9a13-b1a712d77a78 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.505058] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 839.505058] env[62368]: value = "task-1198532" [ 839.505058] env[62368]: _type = "Task" [ 839.505058] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.524650] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc36768-b51c-4455-9bc1-cdba47b14eb2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.530713] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198532, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.804851] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.805109] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.806256] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af27e48a-cf61-4d34-af36-74acab430e6f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.829309] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e75d43e-3520-4cc2-9a07-8f948030c8ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.865645] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Reconfiguring VM to detach interface {{(pid=62368) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 839.866066] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4578423b-5c49-4ce7-ac5e-04c99dbf1ce5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.889259] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 839.889259] env[62368]: value = "task-1198533" [ 839.889259] env[62368]: _type = "Task" [ 839.889259] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.897058] env[62368]: DEBUG nova.compute.utils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.899036] env[62368]: DEBUG nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.899591] env[62368]: DEBUG nova.network.neutron [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 839.910699] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.979926] env[62368]: DEBUG oslo_vmware.api [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198531, 'name': PowerOffVM_Task, 'duration_secs': 0.220535} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.980848] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.981108] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.981415] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1abc4fcc-f58c-4f39-9e37-dafc707e387f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.987731] env[62368]: DEBUG nova.policy [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0800ab273ca04fbf9396175b57eed6b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e821059910b4e32aab596c6f4d521d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.017717] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198532, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.040639] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 840.041087] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d154fb59-681c-48e7-9f38-0f8a7e5c1787 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.051391] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 840.051391] env[62368]: value = "task-1198535" [ 840.051391] env[62368]: _type = "Task" [ 840.051391] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.062533] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198535, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.081256] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.081712] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.082074] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleting the datastore file [datastore2] 29d2f0d6-7a76-4039-acbb-a3abd69d8370 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.082558] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df1ddea6-3805-42cc-9a74-c225d6be049f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.091433] env[62368]: DEBUG oslo_vmware.api [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 840.091433] env[62368]: value = "task-1198536" [ 840.091433] env[62368]: _type = "Task" [ 840.091433] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.102653] env[62368]: DEBUG oslo_vmware.api [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198536, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.208560] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1479f07b-c3f3-4bd9-a6a6-9c9bc5ba59cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.241066] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance 'a0cea538-b162-4504-ac34-803a2d5a8071' progress to 0 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 840.347120] env[62368]: DEBUG nova.network.neutron [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Successfully created port: bd3e0156-36ff-4c90-8484-69fed4dfa149 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.407600] env[62368]: DEBUG nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.411483] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.519915] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198532, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596578} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.520329] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 2a46c954-449e-4d62-be80-add1040ed4c0/2a46c954-449e-4d62-be80-add1040ed4c0.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.520825] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.520825] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4be62546-3dfe-4f6f-a155-c60ebdd1e109 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.531450] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 840.531450] env[62368]: value = "task-1198537" [ 840.531450] env[62368]: _type = "Task" [ 840.531450] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.541262] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198537, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.564306] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198535, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.604104] env[62368]: DEBUG oslo_vmware.api [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198536, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140161} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.606956] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.607058] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.607212] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.607384] env[62368]: INFO nova.compute.manager [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Took 1.16 seconds to destroy the instance on the hypervisor. [ 840.607685] env[62368]: DEBUG oslo.service.loopingcall [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.608039] env[62368]: DEBUG nova.compute.manager [-] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.608145] env[62368]: DEBUG nova.network.neutron [-] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 840.739917] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ac9564-f5eb-4f3b-b2bd-8ce049d1713c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.747436] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 840.750611] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-090da56d-0d70-40c1-8b0a-0e83a7ac1732 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.753946] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3617309-b890-4360-b48d-88ec446f44a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.794212] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0421d5-ce9b-47da-82c6-f871420f1015 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.797198] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 840.797198] env[62368]: value = "task-1198538" [ 840.797198] env[62368]: _type = "Task" [ 840.797198] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.805963] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2263debf-6a48-4b75-8f2a-e4f8c23d4d6e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.812840] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198538, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.823354] env[62368]: DEBUG nova.compute.provider_tree [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.900793] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.947012] env[62368]: DEBUG nova.compute.manager [req-eeaec9b2-5bb4-4040-b733-4fce4d4a7a69 req-fcf7b341-2bc4-44e9-aa4a-b8cbf80e5f24 service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Received event network-vif-deleted-cdba9d6e-63cb-4936-885a-9dc035e3dff0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.947497] env[62368]: INFO nova.compute.manager [req-eeaec9b2-5bb4-4040-b733-4fce4d4a7a69 req-fcf7b341-2bc4-44e9-aa4a-b8cbf80e5f24 service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Neutron deleted interface cdba9d6e-63cb-4936-885a-9dc035e3dff0; detaching it from the instance and deleting it from the info cache [ 840.947741] env[62368]: DEBUG nova.network.neutron [req-eeaec9b2-5bb4-4040-b733-4fce4d4a7a69 req-fcf7b341-2bc4-44e9-aa4a-b8cbf80e5f24 service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.043787] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198537, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.277836} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.044435] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.044983] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d95790-f585-4d16-bcfa-e03fb810ecf5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.067288] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 2a46c954-449e-4d62-be80-add1040ed4c0/2a46c954-449e-4d62-be80-add1040ed4c0.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.070508] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1ea533d-8ff7-408f-8043-bdc82ff8adf9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.091028] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198535, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.092366] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 841.092366] env[62368]: value = "task-1198539" [ 841.092366] env[62368]: _type = "Task" [ 841.092366] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.100974] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198539, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.308848] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198538, 'name': PowerOffVM_Task, 'duration_secs': 0.487521} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.309172] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 841.309395] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance 'a0cea538-b162-4504-ac34-803a2d5a8071' progress to 17 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 841.326363] env[62368]: DEBUG nova.scheduler.client.report [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.401099] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.417403] env[62368]: DEBUG nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.424550] env[62368]: DEBUG nova.network.neutron [-] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.444609] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.444878] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.445050] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.445333] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.445558] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.445720] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.445934] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.446113] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.446287] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.446454] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.446628] env[62368]: DEBUG nova.virt.hardware [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.447714] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb875ea-2ff5-4932-904c-ebb5e72a79ce {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.450908] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e13a322-a2c2-4367-8831-d75956ec6536 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.459477] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf354b5e-3530-46f9-819a-843701593d3a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.466103] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71949a1-739a-4fa5-b7ae-c84aaa8f11d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.496810] env[62368]: DEBUG nova.compute.manager [req-eeaec9b2-5bb4-4040-b733-4fce4d4a7a69 req-fcf7b341-2bc4-44e9-aa4a-b8cbf80e5f24 service nova] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Detach interface failed, port_id=cdba9d6e-63cb-4936-885a-9dc035e3dff0, reason: Instance 29d2f0d6-7a76-4039-acbb-a3abd69d8370 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 841.568278] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198535, 'name': CreateSnapshot_Task, 'duration_secs': 1.262119} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.568568] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 841.569347] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9caa1b5-cc91-46dc-ba34-a825f5001eee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.603309] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198539, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.764485] env[62368]: DEBUG nova.compute.manager [req-2c1d85a9-9e0f-4a91-bfa0-3737c5eb84eb req-58a85195-7d7e-4c96-86bd-f1b1f85095ef service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Received event network-vif-plugged-bd3e0156-36ff-4c90-8484-69fed4dfa149 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.765060] env[62368]: DEBUG oslo_concurrency.lockutils [req-2c1d85a9-9e0f-4a91-bfa0-3737c5eb84eb req-58a85195-7d7e-4c96-86bd-f1b1f85095ef service nova] Acquiring lock "58efc1cf-2469-41be-bf99-fc7b8c72113c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.765329] env[62368]: DEBUG oslo_concurrency.lockutils [req-2c1d85a9-9e0f-4a91-bfa0-3737c5eb84eb req-58a85195-7d7e-4c96-86bd-f1b1f85095ef service nova] Lock "58efc1cf-2469-41be-bf99-fc7b8c72113c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.765735] env[62368]: DEBUG oslo_concurrency.lockutils [req-2c1d85a9-9e0f-4a91-bfa0-3737c5eb84eb req-58a85195-7d7e-4c96-86bd-f1b1f85095ef service nova] Lock "58efc1cf-2469-41be-bf99-fc7b8c72113c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.766154] env[62368]: DEBUG nova.compute.manager [req-2c1d85a9-9e0f-4a91-bfa0-3737c5eb84eb req-58a85195-7d7e-4c96-86bd-f1b1f85095ef service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] No waiting events found dispatching network-vif-plugged-bd3e0156-36ff-4c90-8484-69fed4dfa149 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.766518] env[62368]: WARNING nova.compute.manager [req-2c1d85a9-9e0f-4a91-bfa0-3737c5eb84eb req-58a85195-7d7e-4c96-86bd-f1b1f85095ef service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Received unexpected event network-vif-plugged-bd3e0156-36ff-4c90-8484-69fed4dfa149 for instance with vm_state building and task_state spawning. [ 841.817270] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.817674] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.817916] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.818210] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.818416] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.818599] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.818845] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.819061] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.819288] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.819618] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.819848] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.824950] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d84202c-c4a4-4d5c-bd1f-60ecf63b1926 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.835847] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.836364] env[62368]: DEBUG nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.839521] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.931s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.839879] env[62368]: DEBUG nova.objects.instance [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lazy-loading 'resources' on Instance uuid 54083854-e314-44a7-b4b0-fbef5fa6b1ef {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.847613] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 841.847613] env[62368]: value = "task-1198540" [ 841.847613] env[62368]: _type = "Task" [ 841.847613] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.859287] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198540, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.877967] env[62368]: DEBUG nova.network.neutron [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Successfully updated port: bd3e0156-36ff-4c90-8484-69fed4dfa149 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.902795] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.928095] env[62368]: INFO nova.compute.manager [-] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Took 1.32 seconds to deallocate network for instance. [ 842.087467] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 842.087779] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c56d5d03-b6e3-4e6c-acb4-906effbae3a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.099170] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 842.099170] env[62368]: value = "task-1198541" [ 842.099170] env[62368]: _type = "Task" [ 842.099170] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.106864] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198539, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.112618] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198541, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.346201] env[62368]: DEBUG nova.compute.utils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.347869] env[62368]: DEBUG nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 842.348078] env[62368]: DEBUG nova.network.neutron [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 842.360044] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198540, 'name': ReconfigVM_Task, 'duration_secs': 0.163726} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.360982] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance 'a0cea538-b162-4504-ac34-803a2d5a8071' progress to 33 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 842.380852] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-58efc1cf-2469-41be-bf99-fc7b8c72113c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.381064] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-58efc1cf-2469-41be-bf99-fc7b8c72113c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.381250] env[62368]: DEBUG nova.network.neutron [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 842.393465] env[62368]: DEBUG nova.policy [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7dceb8c8f8b4c4dbdde4d097009d429', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b70fcc6664f47d7b55447210851c4cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 842.411453] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.435521] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.608792] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198539, 'name': ReconfigVM_Task, 'duration_secs': 1.334645} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.608792] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 2a46c954-449e-4d62-be80-add1040ed4c0/2a46c954-449e-4d62-be80-add1040ed4c0.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.608792] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-017da92e-698b-433f-9a12-10c9a477cf20 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.615724] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198541, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.622317] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 842.622317] env[62368]: value = "task-1198542" [ 842.622317] env[62368]: _type = "Task" [ 842.622317] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.631575] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198542, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.655293] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80135c9-8a28-449c-b525-05ba812c4bb3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.662995] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8afb6c2-4965-4307-bf0b-f2c09ea02275 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.698319] env[62368]: DEBUG nova.network.neutron [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Successfully created port: f2865835-685b-4772-8401-30b2455e11c8 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.702057] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b2f9c6-0a69-4e7d-ba1a-39e5b0ecf638 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.710096] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541aa7d1-39c1-4b11-a522-ead7ccf46704 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.725736] env[62368]: DEBUG nova.compute.provider_tree [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.856038] env[62368]: DEBUG nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.870102] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:16:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f181c86a-16f3-4905-b3d9-783b0740eec7',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1115957553',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.870534] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.870761] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.871127] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.871404] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.871678] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.872048] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.872313] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.872714] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.872932] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.873277] env[62368]: DEBUG nova.virt.hardware [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.882410] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfiguring VM instance instance-0000003d to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 842.882410] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-216c5515-ff83-41f6-8e88-495732bfaab0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.912725] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.914434] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 842.914434] env[62368]: value = "task-1198543" [ 842.914434] env[62368]: _type = "Task" [ 842.914434] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.923958] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198543, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.956377] env[62368]: DEBUG nova.network.neutron [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 843.113532] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198541, 'name': CloneVM_Task, 'duration_secs': 0.994457} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.113532] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Created linked-clone VM from snapshot [ 843.113532] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2f08b0-2f6f-4c31-a93a-5547ee7103d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.120318] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Uploading image 5037b35f-25b2-408c-9f87-c33c3892b67e {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 843.133821] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198542, 'name': Rename_Task, 'duration_secs': 0.155919} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.134125] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 843.134402] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73ec952a-e68d-45c0-aaa6-733d3af3f5c3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.142678] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 843.142678] env[62368]: value = "task-1198544" [ 843.142678] env[62368]: _type = "Task" [ 843.142678] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.147574] env[62368]: DEBUG oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 843.147574] env[62368]: value = "vm-259819" [ 843.147574] env[62368]: _type = "VirtualMachine" [ 843.147574] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 843.147828] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2fea05e8-a462-49dc-b252-9bfb9b86a235 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.156855] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198544, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.162964] env[62368]: DEBUG oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease: (returnval){ [ 843.162964] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5288ee69-c739-9ec7-a035-53e1ecbbd655" [ 843.162964] env[62368]: _type = "HttpNfcLease" [ 843.162964] env[62368]: } obtained for exporting VM: (result){ [ 843.162964] env[62368]: value = "vm-259819" [ 843.162964] env[62368]: _type = "VirtualMachine" [ 843.162964] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 843.163250] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the lease: (returnval){ [ 843.163250] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5288ee69-c739-9ec7-a035-53e1ecbbd655" [ 843.163250] env[62368]: _type = "HttpNfcLease" [ 843.163250] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 843.171027] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.171027] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5288ee69-c739-9ec7-a035-53e1ecbbd655" [ 843.171027] env[62368]: _type = "HttpNfcLease" [ 843.171027] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 843.179459] env[62368]: DEBUG nova.network.neutron [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Updating instance_info_cache with network_info: [{"id": "bd3e0156-36ff-4c90-8484-69fed4dfa149", "address": "fa:16:3e:8a:c8:d4", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd3e0156-36", "ovs_interfaceid": "bd3e0156-36ff-4c90-8484-69fed4dfa149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.228920] env[62368]: DEBUG nova.scheduler.client.report [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.411347] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.424021] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198543, 'name': ReconfigVM_Task, 'duration_secs': 0.18935} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.424402] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfigured VM instance instance-0000003d to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 843.425108] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82b4ee1-4c6e-48f1-963a-14505c6788fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.448069] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071/a0cea538-b162-4504-ac34-803a2d5a8071.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.448428] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf462a41-b33c-4332-8fe0-cee97bd2596a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.467931] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 843.467931] env[62368]: value = "task-1198546" [ 843.467931] env[62368]: _type = "Task" [ 843.467931] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.476641] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.655179] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198544, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.672642] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.672642] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5288ee69-c739-9ec7-a035-53e1ecbbd655" [ 843.672642] env[62368]: _type = "HttpNfcLease" [ 843.672642] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 843.672961] env[62368]: DEBUG oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 843.672961] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5288ee69-c739-9ec7-a035-53e1ecbbd655" [ 843.672961] env[62368]: _type = "HttpNfcLease" [ 843.672961] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 843.673704] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1810dc06-4baa-4a25-974a-1a3c7a237c4f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.685150] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-58efc1cf-2469-41be-bf99-fc7b8c72113c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.685481] env[62368]: DEBUG nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Instance network_info: |[{"id": "bd3e0156-36ff-4c90-8484-69fed4dfa149", "address": "fa:16:3e:8a:c8:d4", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd3e0156-36", "ovs_interfaceid": "bd3e0156-36ff-4c90-8484-69fed4dfa149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.685826] env[62368]: DEBUG oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522984b9-625f-0c2f-cb20-a976ee6424e5/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 843.686111] env[62368]: DEBUG oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522984b9-625f-0c2f-cb20-a976ee6424e5/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 843.687633] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:c8:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd3e0156-36ff-4c90-8484-69fed4dfa149', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.697190] env[62368]: DEBUG oslo.service.loopingcall [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.698371] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 843.758402] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b15aa0d9-3257-433f-9e21-07e625ba2aaf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.774574] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.935s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.777879] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.333s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.779457] env[62368]: INFO nova.compute.claims [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.788550] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.788550] env[62368]: value = "task-1198547" [ 843.788550] env[62368]: _type = "Task" [ 843.788550] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.793027] env[62368]: INFO nova.scheduler.client.report [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Deleted allocations for instance 54083854-e314-44a7-b4b0-fbef5fa6b1ef [ 843.797836] env[62368]: DEBUG nova.compute.manager [req-2b41c2fb-d6c4-4d1c-94c3-7936a9f29d21 req-515f2dd2-3b52-4ac8-9aa4-6048527e2dcf service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Received event network-changed-bd3e0156-36ff-4c90-8484-69fed4dfa149 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.798050] env[62368]: DEBUG nova.compute.manager [req-2b41c2fb-d6c4-4d1c-94c3-7936a9f29d21 req-515f2dd2-3b52-4ac8-9aa4-6048527e2dcf service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Refreshing instance network info cache due to event network-changed-bd3e0156-36ff-4c90-8484-69fed4dfa149. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 843.798279] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b41c2fb-d6c4-4d1c-94c3-7936a9f29d21 req-515f2dd2-3b52-4ac8-9aa4-6048527e2dcf service nova] Acquiring lock "refresh_cache-58efc1cf-2469-41be-bf99-fc7b8c72113c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.798431] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b41c2fb-d6c4-4d1c-94c3-7936a9f29d21 req-515f2dd2-3b52-4ac8-9aa4-6048527e2dcf service nova] Acquired lock "refresh_cache-58efc1cf-2469-41be-bf99-fc7b8c72113c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.798594] env[62368]: DEBUG nova.network.neutron [req-2b41c2fb-d6c4-4d1c-94c3-7936a9f29d21 req-515f2dd2-3b52-4ac8-9aa4-6048527e2dcf service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Refreshing network info cache for port bd3e0156-36ff-4c90-8484-69fed4dfa149 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 843.807702] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198547, 'name': CreateVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.824322] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e6631472-6a02-43ce-8460-42596850f44f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.867152] env[62368]: DEBUG nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.890116] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.890437] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.890600] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.890818] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.890998] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.891234] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.891462] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.891628] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.891798] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.891961] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.892262] env[62368]: DEBUG nova.virt.hardware [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.893180] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7f481d-f288-45d6-a295-1cb966c05819 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.904892] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ca5eb6-4295-4c63-ab5d-9f41dca1bd63 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.913929] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.979701] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198546, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.156999] env[62368]: DEBUG oslo_vmware.api [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198544, 'name': PowerOnVM_Task, 'duration_secs': 0.559376} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.157173] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 844.157830] env[62368]: INFO nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Took 9.76 seconds to spawn the instance on the hypervisor. [ 844.158116] env[62368]: DEBUG nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.159725] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d4715f-8f2f-40a2-a6cf-011d3d2fb8af {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.312936] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198547, 'name': CreateVM_Task, 'duration_secs': 0.415281} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.315229] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.315828] env[62368]: DEBUG oslo_concurrency.lockutils [None req-334c92f6-3e81-4fec-9a3f-3af35442cc18 tempest-ServerGroupTestJSON-557713829 tempest-ServerGroupTestJSON-557713829-project-member] Lock "54083854-e314-44a7-b4b0-fbef5fa6b1ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.893s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.321128] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.321128] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.321128] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.321128] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a92e4376-1598-4f6b-8de1-d54e787bf583 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.330595] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 844.330595] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521bdb73-b328-8186-1096-6658c2c929d8" [ 844.330595] env[62368]: _type = "Task" [ 844.330595] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.344240] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521bdb73-b328-8186-1096-6658c2c929d8, 'name': SearchDatastore_Task, 'duration_secs': 0.013019} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.344820] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.345523] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.345805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.346162] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.346510] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.346965] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6502cd82-efcc-48a1-95bd-d557ced58927 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.358634] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.359190] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 844.362886] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a985b073-969c-497d-a107-f8d8cbb0fed2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.370306] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 844.370306] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52d1e4f7-4118-07a2-b4c7-fff0c60e95af" [ 844.370306] env[62368]: _type = "Task" [ 844.370306] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.379798] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d1e4f7-4118-07a2-b4c7-fff0c60e95af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.417454] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.482791] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198546, 'name': ReconfigVM_Task, 'duration_secs': 0.547367} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.482791] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfigured VM instance instance-0000003d to attach disk [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071/a0cea538-b162-4504-ac34-803a2d5a8071.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.482791] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance 'a0cea538-b162-4504-ac34-803a2d5a8071' progress to 50 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 844.519498] env[62368]: DEBUG nova.network.neutron [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Successfully updated port: f2865835-685b-4772-8401-30b2455e11c8 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.621322] env[62368]: DEBUG nova.network.neutron [req-2b41c2fb-d6c4-4d1c-94c3-7936a9f29d21 req-515f2dd2-3b52-4ac8-9aa4-6048527e2dcf service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Updated VIF entry in instance network info cache for port bd3e0156-36ff-4c90-8484-69fed4dfa149. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 844.622050] env[62368]: DEBUG nova.network.neutron [req-2b41c2fb-d6c4-4d1c-94c3-7936a9f29d21 req-515f2dd2-3b52-4ac8-9aa4-6048527e2dcf service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Updating instance_info_cache with network_info: [{"id": "bd3e0156-36ff-4c90-8484-69fed4dfa149", "address": "fa:16:3e:8a:c8:d4", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd3e0156-36", "ovs_interfaceid": "bd3e0156-36ff-4c90-8484-69fed4dfa149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.683685] env[62368]: INFO nova.compute.manager [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Took 43.70 seconds to build instance. [ 844.884301] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d1e4f7-4118-07a2-b4c7-fff0c60e95af, 'name': SearchDatastore_Task, 'duration_secs': 0.010072} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.888380] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7e01fe9-eeae-4c14-afa7-88d5f1bd3c4f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.895714] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 844.895714] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52854d46-595f-4d6f-fc07-e21825dcd742" [ 844.895714] env[62368]: _type = "Task" [ 844.895714] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.910068] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52854d46-595f-4d6f-fc07-e21825dcd742, 'name': SearchDatastore_Task, 'duration_secs': 0.011782} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.910954] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.911450] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 58efc1cf-2469-41be-bf99-fc7b8c72113c/58efc1cf-2469-41be-bf99-fc7b8c72113c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 844.911896] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-942d3593-6dfa-461a-8c1a-0e493c01e619 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.920507] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.927572] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 844.927572] env[62368]: value = "task-1198548" [ 844.927572] env[62368]: _type = "Task" [ 844.927572] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.939437] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198548, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.991740] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8616700-af5b-44d3-8987-fd6a433c876f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.023064] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-5554732b-34a1-41bd-9d0d-9bd39d62bec3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.023219] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-5554732b-34a1-41bd-9d0d-9bd39d62bec3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.025188] env[62368]: DEBUG nova.network.neutron [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.030723] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd56037-764a-42df-9c58-9a1b9b989a3a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.056883] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance 'a0cea538-b162-4504-ac34-803a2d5a8071' progress to 67 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 845.124863] env[62368]: DEBUG oslo_concurrency.lockutils [req-2b41c2fb-d6c4-4d1c-94c3-7936a9f29d21 req-515f2dd2-3b52-4ac8-9aa4-6048527e2dcf service nova] Releasing lock "refresh_cache-58efc1cf-2469-41be-bf99-fc7b8c72113c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.188740] env[62368]: DEBUG oslo_concurrency.lockutils [None req-71d972de-cde6-4a22-83f7-2e8bbb603355 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "2a46c954-449e-4d62-be80-add1040ed4c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.675s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.215834] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ab7de1-5478-4fc9-8994-f6d30bb42663 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.227532] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17473f5f-3e04-4e71-8202-60c3e4efbee1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.272957] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ce1d3b-5c73-4f3a-a765-081da3acea8f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.283658] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72440219-dc15-4dae-8876-c5ea1bb3194a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.304548] env[62368]: DEBUG nova.compute.provider_tree [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.414643] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.441171] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198548, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.592085] env[62368]: DEBUG nova.network.neutron [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.688301] env[62368]: DEBUG nova.network.neutron [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Port 444a47e3-2f56-4895-8a8b-7e5605dbfaf1 binding to destination host cpu-1 is already ACTIVE {{(pid=62368) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 845.809265] env[62368]: DEBUG nova.scheduler.client.report [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.837992] env[62368]: DEBUG nova.compute.manager [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Received event network-vif-plugged-f2865835-685b-4772-8401-30b2455e11c8 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.838412] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Acquiring lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.838767] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.839256] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.840697] env[62368]: DEBUG nova.compute.manager [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] No waiting events found dispatching network-vif-plugged-f2865835-685b-4772-8401-30b2455e11c8 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 845.840697] env[62368]: WARNING nova.compute.manager [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Received unexpected event network-vif-plugged-f2865835-685b-4772-8401-30b2455e11c8 for instance with vm_state building and task_state spawning. [ 845.840697] env[62368]: DEBUG nova.compute.manager [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Received event network-changed-f2865835-685b-4772-8401-30b2455e11c8 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.840697] env[62368]: DEBUG nova.compute.manager [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Refreshing instance network info cache due to event network-changed-f2865835-685b-4772-8401-30b2455e11c8. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.840697] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Acquiring lock "refresh_cache-5554732b-34a1-41bd-9d0d-9bd39d62bec3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.898279] env[62368]: DEBUG nova.network.neutron [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Updating instance_info_cache with network_info: [{"id": "f2865835-685b-4772-8401-30b2455e11c8", "address": "fa:16:3e:48:44:61", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2865835-68", "ovs_interfaceid": "f2865835-685b-4772-8401-30b2455e11c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.913975] env[62368]: DEBUG oslo_vmware.api [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198533, 'name': ReconfigVM_Task, 'duration_secs': 5.860166} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.914500] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.914561] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Reconfigured VM to detach interface {{(pid=62368) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 845.941358] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198548, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518201} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.941649] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 58efc1cf-2469-41be-bf99-fc7b8c72113c/58efc1cf-2469-41be-bf99-fc7b8c72113c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 845.942124] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.942489] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1127b004-2421-41e8-b796-9f758d314680 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.951073] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 845.951073] env[62368]: value = "task-1198549" [ 845.951073] env[62368]: _type = "Task" [ 845.951073] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.965085] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.318402] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.318951] env[62368]: DEBUG nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.323280] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.768s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.325674] env[62368]: INFO nova.compute.claims [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.406182] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-5554732b-34a1-41bd-9d0d-9bd39d62bec3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.406182] env[62368]: DEBUG nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Instance network_info: |[{"id": "f2865835-685b-4772-8401-30b2455e11c8", "address": "fa:16:3e:48:44:61", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2865835-68", "ovs_interfaceid": "f2865835-685b-4772-8401-30b2455e11c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.406182] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Acquired lock "refresh_cache-5554732b-34a1-41bd-9d0d-9bd39d62bec3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.406182] env[62368]: DEBUG nova.network.neutron [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Refreshing network info cache for port f2865835-685b-4772-8401-30b2455e11c8 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.407688] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:44:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2865835-685b-4772-8401-30b2455e11c8', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.418989] env[62368]: DEBUG oslo.service.loopingcall [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.419920] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 846.421675] env[62368]: DEBUG nova.compute.manager [req-73a1a06a-e4d5-407e-a41d-43b6b0fb4436 req-0074e747-6a1e-4ac7-96d7-aabf488dbbd2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-vif-deleted-90ff566e-24b4-4a83-a965-0bc25b1152e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.422040] env[62368]: INFO nova.compute.manager [req-73a1a06a-e4d5-407e-a41d-43b6b0fb4436 req-0074e747-6a1e-4ac7-96d7-aabf488dbbd2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Neutron deleted interface 90ff566e-24b4-4a83-a965-0bc25b1152e0; detaching it from the instance and deleting it from the info cache [ 846.422495] env[62368]: DEBUG nova.network.neutron [req-73a1a06a-e4d5-407e-a41d-43b6b0fb4436 req-0074e747-6a1e-4ac7-96d7-aabf488dbbd2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c582c777-e544-40af-9f04-d8c171475bc1", "address": "fa:16:3e:e4:28:b5", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc582c777-e5", "ovs_interfaceid": "c582c777-e544-40af-9f04-d8c171475bc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.426282] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a61c3d69-89a3-42bf-89bd-5e63a4f631ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.442910] env[62368]: DEBUG oslo_concurrency.lockutils [req-73a1a06a-e4d5-407e-a41d-43b6b0fb4436 req-0074e747-6a1e-4ac7-96d7-aabf488dbbd2 service nova] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.443391] env[62368]: DEBUG oslo_concurrency.lockutils [req-73a1a06a-e4d5-407e-a41d-43b6b0fb4436 req-0074e747-6a1e-4ac7-96d7-aabf488dbbd2 service nova] Acquired lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.445039] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4565fe1-1c91-4c2a-a773-0774a82c9ca2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.467651] env[62368]: DEBUG oslo_concurrency.lockutils [req-73a1a06a-e4d5-407e-a41d-43b6b0fb4436 req-0074e747-6a1e-4ac7-96d7-aabf488dbbd2 service nova] Releasing lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.468157] env[62368]: WARNING nova.compute.manager [req-73a1a06a-e4d5-407e-a41d-43b6b0fb4436 req-0074e747-6a1e-4ac7-96d7-aabf488dbbd2 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Detach interface failed, port_id=90ff566e-24b4-4a83-a965-0bc25b1152e0, reason: No device with interface-id 90ff566e-24b4-4a83-a965-0bc25b1152e0 exists on VM: nova.exception.NotFound: No device with interface-id 90ff566e-24b4-4a83-a965-0bc25b1152e0 exists on VM [ 846.475418] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.475418] env[62368]: value = "task-1198550" [ 846.475418] env[62368]: _type = "Task" [ 846.475418] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.484281] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068352} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.485069] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.486009] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acb5bb5-f9f4-4b2f-9ca6-9074a37b715c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.491788] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198550, 'name': CreateVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.516157] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 58efc1cf-2469-41be-bf99-fc7b8c72113c/58efc1cf-2469-41be-bf99-fc7b8c72113c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.516920] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b3fbedf-5df9-46ec-b8ee-c43c532931ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.540382] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 846.540382] env[62368]: value = "task-1198551" [ 846.540382] env[62368]: _type = "Task" [ 846.540382] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.549027] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198551, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.713044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.713313] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.713481] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.832232] env[62368]: DEBUG nova.compute.utils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.834602] env[62368]: DEBUG nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.838026] env[62368]: DEBUG nova.network.neutron [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 846.889779] env[62368]: DEBUG nova.policy [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eaedf6f5fa314485882377378995c30e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1fafa1daaec346e7b6a90923c6c30285', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.989429] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198550, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.056537] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198551, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.215100] env[62368]: DEBUG nova.network.neutron [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Successfully created port: 932cae7b-da6e-406b-8d5a-a5e049d2cffd {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.284137] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.284355] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.284562] env[62368]: DEBUG nova.network.neutron [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.341349] env[62368]: DEBUG nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.405850] env[62368]: DEBUG nova.network.neutron [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Updated VIF entry in instance network info cache for port f2865835-685b-4772-8401-30b2455e11c8. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 847.405850] env[62368]: DEBUG nova.network.neutron [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Updating instance_info_cache with network_info: [{"id": "f2865835-685b-4772-8401-30b2455e11c8", "address": "fa:16:3e:48:44:61", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2865835-68", "ovs_interfaceid": "f2865835-685b-4772-8401-30b2455e11c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.488566] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198550, 'name': CreateVM_Task, 'duration_secs': 1.011509} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.491163] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 847.492435] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.492614] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.493702] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.493702] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ce6da52-f8bd-4ded-beba-75e679a10239 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.499284] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 847.499284] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]523aec23-51bc-a759-264d-90bc49969a49" [ 847.499284] env[62368]: _type = "Task" [ 847.499284] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.509897] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523aec23-51bc-a759-264d-90bc49969a49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.553271] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198551, 'name': ReconfigVM_Task, 'duration_secs': 0.672381} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.553558] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 58efc1cf-2469-41be-bf99-fc7b8c72113c/58efc1cf-2469-41be-bf99-fc7b8c72113c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.554670] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50ad66bd-9c07-447b-a958-88ea610aa32f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.562544] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 847.562544] env[62368]: value = "task-1198552" [ 847.562544] env[62368]: _type = "Task" [ 847.562544] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.574826] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198552, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.710596] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bedb3a4-80fb-4cf6-b7a6-5a6e115b84d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.724755] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60386a63-c487-4088-b392-1dd17a3116ab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.766216] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e03d841-a2d9-4317-b499-aff687beb8c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.775669] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c61516-424a-49bc-a6fd-b297ad1a70f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.794673] env[62368]: DEBUG nova.compute.provider_tree [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.797281] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.797418] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.797577] env[62368]: DEBUG nova.network.neutron [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.907205] env[62368]: DEBUG nova.compute.manager [req-5dcb85ab-58b6-464e-ab96-00f27212f8a8 req-4757df3f-68c3-473e-a4d1-200144a83bcf service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-vif-deleted-c582c777-e544-40af-9f04-d8c171475bc1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.907205] env[62368]: INFO nova.compute.manager [req-5dcb85ab-58b6-464e-ab96-00f27212f8a8 req-4757df3f-68c3-473e-a4d1-200144a83bcf service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Neutron deleted interface c582c777-e544-40af-9f04-d8c171475bc1; detaching it from the instance and deleting it from the info cache [ 847.907205] env[62368]: DEBUG nova.network.neutron [req-5dcb85ab-58b6-464e-ab96-00f27212f8a8 req-4757df3f-68c3-473e-a4d1-200144a83bcf service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.913371] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Releasing lock "refresh_cache-5554732b-34a1-41bd-9d0d-9bd39d62bec3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.913698] env[62368]: DEBUG nova.compute.manager [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Received event network-changed-b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.913828] env[62368]: DEBUG nova.compute.manager [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Refreshing instance network info cache due to event network-changed-b4fab9b2-299a-4e59-92bf-efdc3d583416. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.914050] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Acquiring lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.914709] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Acquired lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.914709] env[62368]: DEBUG nova.network.neutron [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Refreshing network info cache for port b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 847.958939] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.959246] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.959517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.959641] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.959815] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.962135] env[62368]: INFO nova.compute.manager [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Terminating instance [ 847.964111] env[62368]: DEBUG nova.compute.manager [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 847.964332] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 847.965195] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed85969-49f7-420e-b673-c2523e71e249 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.974553] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 847.974937] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1de80f63-118f-4f3a-ad13-d23f2991232a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.982073] env[62368]: DEBUG oslo_vmware.api [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 847.982073] env[62368]: value = "task-1198553" [ 847.982073] env[62368]: _type = "Task" [ 847.982073] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.991732] env[62368]: DEBUG oslo_vmware.api [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198553, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.013359] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523aec23-51bc-a759-264d-90bc49969a49, 'name': SearchDatastore_Task, 'duration_secs': 0.010633} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.013864] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.014413] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.014647] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.014813] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.015280] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.015594] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26f27a9b-3438-4e72-9993-1772092c3011 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.026653] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.026880] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 848.027667] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fd1416c-f5ba-4654-8ed2-683cb59200d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.035178] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 848.035178] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52baa780-a0b2-f5b9-7d10-d89a15f4e7ff" [ 848.035178] env[62368]: _type = "Task" [ 848.035178] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.046474] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52baa780-a0b2-f5b9-7d10-d89a15f4e7ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.073297] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198552, 'name': Rename_Task, 'duration_secs': 0.230936} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.073593] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 848.073858] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28293dee-6921-42b2-9624-cd61d694b13c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.081812] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 848.081812] env[62368]: value = "task-1198554" [ 848.081812] env[62368]: _type = "Task" [ 848.081812] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.091371] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.124850] env[62368]: INFO nova.network.neutron [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Port c582c777-e544-40af-9f04-d8c171475bc1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 848.125538] env[62368]: DEBUG nova.network.neutron [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [{"id": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "address": "fa:16:3e:64:4e:49", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a1f27ab-91", "ovs_interfaceid": "6a1f27ab-91ba-4185-9fdb-c5e31d622ed6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.297655] env[62368]: DEBUG nova.scheduler.client.report [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.353764] env[62368]: DEBUG nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.379645] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.379940] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.380117] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.380303] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.380450] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.380704] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.380953] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.381132] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.381304] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.381470] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.381650] env[62368]: DEBUG nova.virt.hardware [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.382592] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791636f5-587f-4705-8c80-274809a35d6e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.393478] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9f449c-bf09-4249-b765-8ec8eef6e5ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.409969] env[62368]: DEBUG oslo_concurrency.lockutils [req-5dcb85ab-58b6-464e-ab96-00f27212f8a8 req-4757df3f-68c3-473e-a4d1-200144a83bcf service nova] Acquiring lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.494683] env[62368]: DEBUG oslo_vmware.api [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198553, 'name': PowerOffVM_Task, 'duration_secs': 0.294084} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.494683] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 848.494683] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 848.494683] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed0e0297-8a16-469f-8976-8392e2b66e7d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.511643] env[62368]: DEBUG nova.network.neutron [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance_info_cache with network_info: [{"id": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "address": "fa:16:3e:98:a0:d2", "network": {"id": "34f156be-cab3-45a2-8b52-74dd4cd1e42d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2a054e975862463bb842479ad19393ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap444a47e3-2f", "ovs_interfaceid": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.549071] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52baa780-a0b2-f5b9-7d10-d89a15f4e7ff, 'name': SearchDatastore_Task, 'duration_secs': 0.010377} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.549992] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f278f7f-9bbd-4086-b0ab-658d5b33c980 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.556496] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 848.556496] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52a9c2fb-fcc4-5d3b-4209-00aaa0426863" [ 848.556496] env[62368]: _type = "Task" [ 848.556496] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.567633] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a9c2fb-fcc4-5d3b-4209-00aaa0426863, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.591701] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198554, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.629221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.697710] env[62368]: DEBUG nova.network.neutron [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updated VIF entry in instance network info cache for port b4fab9b2-299a-4e59-92bf-efdc3d583416. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 848.698115] env[62368]: DEBUG nova.network.neutron [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updating instance_info_cache with network_info: [{"id": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "address": "fa:16:3e:2d:06:3b", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4fab9b2-29", "ovs_interfaceid": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.805278] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.806278] env[62368]: DEBUG nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.809339] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.061s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.811649] env[62368]: INFO nova.compute.claims [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.897845] env[62368]: DEBUG nova.compute.manager [req-e9777b35-e143-4093-a6f8-9a4af33e4e24 req-9727d96a-48d4-4552-9738-072da2e8ec2b service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Received event network-vif-plugged-932cae7b-da6e-406b-8d5a-a5e049d2cffd {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.897845] env[62368]: DEBUG oslo_concurrency.lockutils [req-e9777b35-e143-4093-a6f8-9a4af33e4e24 req-9727d96a-48d4-4552-9738-072da2e8ec2b service nova] Acquiring lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.898040] env[62368]: DEBUG oslo_concurrency.lockutils [req-e9777b35-e143-4093-a6f8-9a4af33e4e24 req-9727d96a-48d4-4552-9738-072da2e8ec2b service nova] Lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.898234] env[62368]: DEBUG oslo_concurrency.lockutils [req-e9777b35-e143-4093-a6f8-9a4af33e4e24 req-9727d96a-48d4-4552-9738-072da2e8ec2b service nova] Lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.898410] env[62368]: DEBUG nova.compute.manager [req-e9777b35-e143-4093-a6f8-9a4af33e4e24 req-9727d96a-48d4-4552-9738-072da2e8ec2b service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] No waiting events found dispatching network-vif-plugged-932cae7b-da6e-406b-8d5a-a5e049d2cffd {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.898582] env[62368]: WARNING nova.compute.manager [req-e9777b35-e143-4093-a6f8-9a4af33e4e24 req-9727d96a-48d4-4552-9738-072da2e8ec2b service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Received unexpected event network-vif-plugged-932cae7b-da6e-406b-8d5a-a5e049d2cffd for instance with vm_state building and task_state spawning. [ 848.902150] env[62368]: DEBUG nova.network.neutron [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Successfully updated port: 932cae7b-da6e-406b-8d5a-a5e049d2cffd {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.015047] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.068242] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a9c2fb-fcc4-5d3b-4209-00aaa0426863, 'name': SearchDatastore_Task, 'duration_secs': 0.011404} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.069256] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.069517] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 5554732b-34a1-41bd-9d0d-9bd39d62bec3/5554732b-34a1-41bd-9d0d-9bd39d62bec3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 849.069775] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f139dfd-8e2f-4592-9001-c615c4a73dbb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.078668] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 849.078668] env[62368]: value = "task-1198556" [ 849.078668] env[62368]: _type = "Task" [ 849.078668] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.089802] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.096051] env[62368]: DEBUG oslo_vmware.api [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198554, 'name': PowerOnVM_Task, 'duration_secs': 0.752692} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.096051] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 849.096051] env[62368]: INFO nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Took 7.68 seconds to spawn the instance on the hypervisor. [ 849.096299] env[62368]: DEBUG nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.097035] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa078ec2-b719-4548-9c0f-b25b6513a80e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.133972] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7f5dbac9-2f8c-4c2f-9cb8-868a77d5c15f tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-67d8c43a-d3af-407b-847a-1c2c8c53dbfb-90ff566e-24b4-4a83-a965-0bc25b1152e0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.832s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.201061] env[62368]: DEBUG oslo_concurrency.lockutils [req-b3ca5fcd-e9c4-4d40-a46c-f4c848be5ef3 req-3a96300a-e50f-4b89-9a55-dda198a841bd service nova] Releasing lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.316146] env[62368]: DEBUG nova.compute.utils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.319834] env[62368]: DEBUG nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.320114] env[62368]: DEBUG nova.network.neutron [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 849.360189] env[62368]: DEBUG nova.policy [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c66c425264424e3eaa37c8058e62cb30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '026a3176abc34001ab3506445226777f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.404955] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquiring lock "refresh_cache-044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.405259] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquired lock "refresh_cache-044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.405349] env[62368]: DEBUG nova.network.neutron [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 849.541185] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43eb3f0c-12a4-4963-be1f-fedf3f08e052 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.561721] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f403d50d-91a2-4f6d-b076-347161d1e10b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.569177] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance 'a0cea538-b162-4504-ac34-803a2d5a8071' progress to 83 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 849.588253] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198556, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474631} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.588502] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 5554732b-34a1-41bd-9d0d-9bd39d62bec3/5554732b-34a1-41bd-9d0d-9bd39d62bec3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 849.588736] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.588963] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-05b93db2-11d2-4122-acba-64baee96eb9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.595845] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 849.595845] env[62368]: value = "task-1198557" [ 849.595845] env[62368]: _type = "Task" [ 849.595845] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.604262] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.616805] env[62368]: INFO nova.compute.manager [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Took 44.27 seconds to build instance. [ 849.645968] env[62368]: DEBUG nova.network.neutron [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Successfully created port: b6d50c17-ef64-4171-8459-4829ffa5d03d {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.823040] env[62368]: DEBUG nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.961113] env[62368]: DEBUG nova.network.neutron [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.076705] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 850.077329] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64f7ea3e-b528-4ab4-8ee7-324aa1408b50 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.089433] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 850.089433] env[62368]: value = "task-1198558" [ 850.089433] env[62368]: _type = "Task" [ 850.089433] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.115551] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.13761} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.115835] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198558, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.119239] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.119992] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b92c69b9-b0a0-4cc3-9ab4-b1734897aec5 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "58efc1cf-2469-41be-bf99-fc7b8c72113c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.757s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.123504] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9b8785-c894-4254-8276-2967639ea5b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.151019] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 5554732b-34a1-41bd-9d0d-9bd39d62bec3/5554732b-34a1-41bd-9d0d-9bd39d62bec3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.155118] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d532c93-dd63-4b71-8a34-05a3e9aac68b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.174270] env[62368]: DEBUG nova.network.neutron [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Updating instance_info_cache with network_info: [{"id": "932cae7b-da6e-406b-8d5a-a5e049d2cffd", "address": "fa:16:3e:b2:44:26", "network": {"id": "f469820c-f81c-47cf-99f5-fdbf3a4205c4", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-26036653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1fafa1daaec346e7b6a90923c6c30285", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52c1f5eb-3d4a-4faa-a30d-2b0a46430791", "external-id": "nsx-vlan-transportzone-775", "segmentation_id": 775, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932cae7b-da", "ovs_interfaceid": "932cae7b-da6e-406b-8d5a-a5e049d2cffd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.185063] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 850.185063] env[62368]: value = "task-1198559" [ 850.185063] env[62368]: _type = "Task" [ 850.185063] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.199042] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198559, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.255976] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df54b86a-d62f-4f48-99d3-f0abec538e42 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.265027] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7cb647-856f-4304-b821-106d47c1c70a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.299859] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418f407c-9145-4112-9678-475eed58fc8e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.304597] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 850.304597] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 850.304597] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleting the datastore file [datastore1] 67d8c43a-d3af-407b-847a-1c2c8c53dbfb {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.304597] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9714c34f-aa2a-4e46-b36a-0a9ee77d9c35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.310780] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a020063-d8ac-4e1c-b03b-2edad076575b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.315598] env[62368]: DEBUG oslo_vmware.api [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 850.315598] env[62368]: value = "task-1198560" [ 850.315598] env[62368]: _type = "Task" [ 850.315598] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.327595] env[62368]: DEBUG nova.compute.provider_tree [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.339963] env[62368]: DEBUG oslo_vmware.api [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198560, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.602864] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198558, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.678355] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Releasing lock "refresh_cache-044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.678737] env[62368]: DEBUG nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Instance network_info: |[{"id": "932cae7b-da6e-406b-8d5a-a5e049d2cffd", "address": "fa:16:3e:b2:44:26", "network": {"id": "f469820c-f81c-47cf-99f5-fdbf3a4205c4", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-26036653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1fafa1daaec346e7b6a90923c6c30285", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52c1f5eb-3d4a-4faa-a30d-2b0a46430791", "external-id": "nsx-vlan-transportzone-775", "segmentation_id": 775, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932cae7b-da", "ovs_interfaceid": "932cae7b-da6e-406b-8d5a-a5e049d2cffd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.679211] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:44:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '52c1f5eb-3d4a-4faa-a30d-2b0a46430791', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '932cae7b-da6e-406b-8d5a-a5e049d2cffd', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.686856] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Creating folder: Project (1fafa1daaec346e7b6a90923c6c30285). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 850.687229] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2fd9359b-fba0-4e00-a1f9-5cf36b401fee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.698859] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198559, 'name': ReconfigVM_Task, 'duration_secs': 0.359099} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.699199] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 5554732b-34a1-41bd-9d0d-9bd39d62bec3/5554732b-34a1-41bd-9d0d-9bd39d62bec3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.699914] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c920f1c2-350c-4d53-9069-4a572cd157b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.703112] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Created folder: Project (1fafa1daaec346e7b6a90923c6c30285) in parent group-v259706. [ 850.703547] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Creating folder: Instances. Parent ref: group-v259822. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 850.704140] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9551a8bc-7383-4ffa-86d2-19ac63a38ac9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.707647] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 850.707647] env[62368]: value = "task-1198562" [ 850.707647] env[62368]: _type = "Task" [ 850.707647] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.716112] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198562, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.717439] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Created folder: Instances in parent group-v259822. [ 850.717565] env[62368]: DEBUG oslo.service.loopingcall [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.717718] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 850.717924] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c83330b-6d5c-44df-aa6e-e799bbfdd452 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.738548] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.738548] env[62368]: value = "task-1198564" [ 850.738548] env[62368]: _type = "Task" [ 850.738548] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.747063] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198564, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.829888] env[62368]: DEBUG oslo_vmware.api [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198560, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224546} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.830332] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.830558] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.830813] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.831157] env[62368]: INFO nova.compute.manager [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Took 2.87 seconds to destroy the instance on the hypervisor. [ 850.831460] env[62368]: DEBUG oslo.service.loopingcall [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.831760] env[62368]: DEBUG nova.compute.manager [-] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.831919] env[62368]: DEBUG nova.network.neutron [-] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.835351] env[62368]: DEBUG nova.scheduler.client.report [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.842046] env[62368]: DEBUG nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.876658] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.877158] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.877281] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.877473] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.877624] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.877775] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.877986] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.878171] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.878344] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.878518] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.878696] env[62368]: DEBUG nova.virt.hardware [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.879797] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0daf98-dbab-477e-a1be-e2786f62cb6f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.888589] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f84e5a-955e-4b85-9e08-601af9eb10d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.922759] env[62368]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port c582c777-e544-40af-9f04-d8c171475bc1 could not be found.", "detail": ""}} {{(pid=62368) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 850.923107] env[62368]: DEBUG nova.network.neutron [-] Unable to show port c582c777-e544-40af-9f04-d8c171475bc1 as it no longer exists. {{(pid=62368) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 850.968332] env[62368]: DEBUG nova.compute.manager [req-ecdb9c16-62e0-46e7-b2aa-497e75ea6f5f req-9e264be6-292e-450e-8a4a-0a352467c543 service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Received event network-changed-932cae7b-da6e-406b-8d5a-a5e049d2cffd {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.968515] env[62368]: DEBUG nova.compute.manager [req-ecdb9c16-62e0-46e7-b2aa-497e75ea6f5f req-9e264be6-292e-450e-8a4a-0a352467c543 service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Refreshing instance network info cache due to event network-changed-932cae7b-da6e-406b-8d5a-a5e049d2cffd. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.968736] env[62368]: DEBUG oslo_concurrency.lockutils [req-ecdb9c16-62e0-46e7-b2aa-497e75ea6f5f req-9e264be6-292e-450e-8a4a-0a352467c543 service nova] Acquiring lock "refresh_cache-044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.968884] env[62368]: DEBUG oslo_concurrency.lockutils [req-ecdb9c16-62e0-46e7-b2aa-497e75ea6f5f req-9e264be6-292e-450e-8a4a-0a352467c543 service nova] Acquired lock "refresh_cache-044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.969059] env[62368]: DEBUG nova.network.neutron [req-ecdb9c16-62e0-46e7-b2aa-497e75ea6f5f req-9e264be6-292e-450e-8a4a-0a352467c543 service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Refreshing network info cache for port 932cae7b-da6e-406b-8d5a-a5e049d2cffd {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 851.104885] env[62368]: DEBUG oslo_vmware.api [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198558, 'name': PowerOnVM_Task, 'duration_secs': 0.529654} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.105249] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 851.105522] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9019f92a-b7ac-4ac5-8487-2823b01caefb tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance 'a0cea538-b162-4504-ac34-803a2d5a8071' progress to 100 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 851.217604] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198562, 'name': Rename_Task, 'duration_secs': 0.196873} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.217922] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 851.218205] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e24510c-f174-496a-81a9-c5a88f6476da {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.225276] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 851.225276] env[62368]: value = "task-1198565" [ 851.225276] env[62368]: _type = "Task" [ 851.225276] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.236257] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198565, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.248615] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198564, 'name': CreateVM_Task, 'duration_secs': 0.367898} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.248920] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 851.250758] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.250758] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.250758] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.250758] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cb2be2f-a6fb-4aaa-9baa-ae33f05218a5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.255959] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 851.255959] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522b6bc4-6d10-9063-e243-2f8eb30d07f2" [ 851.255959] env[62368]: _type = "Task" [ 851.255959] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.267644] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522b6bc4-6d10-9063-e243-2f8eb30d07f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.347053] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.347053] env[62368]: DEBUG nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.347274] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.721s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.348711] env[62368]: INFO nova.compute.claims [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.497404] env[62368]: DEBUG nova.network.neutron [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Successfully updated port: b6d50c17-ef64-4171-8459-4829ffa5d03d {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.692632] env[62368]: DEBUG nova.compute.manager [req-c7ea2e10-07b6-4006-8a2e-d817ab490e08 req-36d34f6a-5e6e-4095-a332-6cd31fac0934 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Received event network-vif-deleted-6a1f27ab-91ba-4185-9fdb-c5e31d622ed6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.692888] env[62368]: INFO nova.compute.manager [req-c7ea2e10-07b6-4006-8a2e-d817ab490e08 req-36d34f6a-5e6e-4095-a332-6cd31fac0934 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Neutron deleted interface 6a1f27ab-91ba-4185-9fdb-c5e31d622ed6; detaching it from the instance and deleting it from the info cache [ 851.693104] env[62368]: DEBUG nova.network.neutron [req-c7ea2e10-07b6-4006-8a2e-d817ab490e08 req-36d34f6a-5e6e-4095-a332-6cd31fac0934 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.742980] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198565, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.768662] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522b6bc4-6d10-9063-e243-2f8eb30d07f2, 'name': SearchDatastore_Task, 'duration_secs': 0.012502} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.768662] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.768662] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.768662] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.768662] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.769021] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.769461] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33956c8c-a7e9-4df3-a8f2-558f60b7e133 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.779836] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.780060] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 851.781251] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9ee2602-3767-473f-946d-63a22d16ee62 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.790355] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 851.790355] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522908bd-5959-570c-087a-9a1648f443eb" [ 851.790355] env[62368]: _type = "Task" [ 851.790355] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.801236] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522908bd-5959-570c-087a-9a1648f443eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.849213] env[62368]: DEBUG nova.network.neutron [req-ecdb9c16-62e0-46e7-b2aa-497e75ea6f5f req-9e264be6-292e-450e-8a4a-0a352467c543 service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Updated VIF entry in instance network info cache for port 932cae7b-da6e-406b-8d5a-a5e049d2cffd. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 851.849617] env[62368]: DEBUG nova.network.neutron [req-ecdb9c16-62e0-46e7-b2aa-497e75ea6f5f req-9e264be6-292e-450e-8a4a-0a352467c543 service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Updating instance_info_cache with network_info: [{"id": "932cae7b-da6e-406b-8d5a-a5e049d2cffd", "address": "fa:16:3e:b2:44:26", "network": {"id": "f469820c-f81c-47cf-99f5-fdbf3a4205c4", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-26036653-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1fafa1daaec346e7b6a90923c6c30285", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52c1f5eb-3d4a-4faa-a30d-2b0a46430791", "external-id": "nsx-vlan-transportzone-775", "segmentation_id": 775, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932cae7b-da", "ovs_interfaceid": "932cae7b-da6e-406b-8d5a-a5e049d2cffd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.855716] env[62368]: DEBUG nova.compute.utils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.859019] env[62368]: DEBUG nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.859200] env[62368]: DEBUG nova.network.neutron [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 851.906046] env[62368]: DEBUG nova.policy [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc2d082d49ba4dd480535c2c56501fff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e250606f3df240eb93ce45925b00ad43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.000072] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "refresh_cache-5ed9cee5-8cc8-4b70-96a9-737019363638" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.000236] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "refresh_cache-5ed9cee5-8cc8-4b70-96a9-737019363638" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.000771] env[62368]: DEBUG nova.network.neutron [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.032925] env[62368]: INFO nova.compute.manager [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Rebuilding instance [ 852.094035] env[62368]: DEBUG nova.compute.manager [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.095050] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70d04ea-d7bd-4e50-a9aa-0e2d780aeb6f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.139700] env[62368]: DEBUG nova.network.neutron [-] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.199729] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1610a81-4ac4-403c-900a-a6c1281ae341 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.214149] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337b8240-e89e-4731-b110-9c96e2f93465 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.235661] env[62368]: DEBUG oslo_vmware.api [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198565, 'name': PowerOnVM_Task, 'duration_secs': 0.57366} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.235661] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 852.235661] env[62368]: INFO nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Took 8.37 seconds to spawn the instance on the hypervisor. [ 852.235864] env[62368]: DEBUG nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.247163] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bd2029-bca8-48c1-8801-4f67c2187d6c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.249175] env[62368]: DEBUG nova.compute.manager [req-c7ea2e10-07b6-4006-8a2e-d817ab490e08 req-36d34f6a-5e6e-4095-a332-6cd31fac0934 service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Detach interface failed, port_id=6a1f27ab-91ba-4185-9fdb-c5e31d622ed6, reason: Instance 67d8c43a-d3af-407b-847a-1c2c8c53dbfb could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 852.250279] env[62368]: DEBUG nova.network.neutron [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Successfully created port: d6d93849-79aa-4366-a8b1-ce1732983fa4 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.302903] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522908bd-5959-570c-087a-9a1648f443eb, 'name': SearchDatastore_Task, 'duration_secs': 0.012863} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.303771] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04c2069e-1594-4650-b157-6e343b0d4717 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.309844] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 852.309844] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ce0dea-32f2-d004-b1f3-e6129a516c75" [ 852.309844] env[62368]: _type = "Task" [ 852.309844] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.322160] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ce0dea-32f2-d004-b1f3-e6129a516c75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.354233] env[62368]: DEBUG oslo_concurrency.lockutils [req-ecdb9c16-62e0-46e7-b2aa-497e75ea6f5f req-9e264be6-292e-450e-8a4a-0a352467c543 service nova] Releasing lock "refresh_cache-044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.365437] env[62368]: DEBUG nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.549951] env[62368]: DEBUG nova.network.neutron [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.609944] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 852.610282] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d9a8c40-6f6e-4de8-90b1-d41e2eb6129e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.627238] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 852.627238] env[62368]: value = "task-1198566" [ 852.627238] env[62368]: _type = "Task" [ 852.627238] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.646542] env[62368]: INFO nova.compute.manager [-] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Took 1.81 seconds to deallocate network for instance. [ 852.647076] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198566, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.703809] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6072beb8-c55b-4c4b-a7f1-3ebede9228fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.714026] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e9b289-d824-4fa5-af0d-16d44da4e9a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.747181] env[62368]: DEBUG nova.network.neutron [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Updating instance_info_cache with network_info: [{"id": "b6d50c17-ef64-4171-8459-4829ffa5d03d", "address": "fa:16:3e:7d:ed:a7", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6d50c17-ef", "ovs_interfaceid": "b6d50c17-ef64-4171-8459-4829ffa5d03d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.749948] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776b9f79-9bde-411d-86ef-0dbdf5d7c8c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.758685] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff77f6e-0043-4578-84bc-d3bb7aaf6724 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.771021] env[62368]: INFO nova.compute.manager [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Took 44.99 seconds to build instance. [ 852.780233] env[62368]: DEBUG nova.compute.provider_tree [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.821101] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ce0dea-32f2-d004-b1f3-e6129a516c75, 'name': SearchDatastore_Task, 'duration_secs': 0.029139} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.821467] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.821828] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5/044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 852.822192] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b36dbb84-30c9-4531-8928-2a9491095968 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.830051] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 852.830051] env[62368]: value = "task-1198567" [ 852.830051] env[62368]: _type = "Task" [ 852.830051] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.839235] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.996457] env[62368]: DEBUG nova.compute.manager [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Received event network-vif-plugged-b6d50c17-ef64-4171-8459-4829ffa5d03d {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.996727] env[62368]: DEBUG oslo_concurrency.lockutils [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] Acquiring lock "5ed9cee5-8cc8-4b70-96a9-737019363638-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.996925] env[62368]: DEBUG oslo_concurrency.lockutils [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] Lock "5ed9cee5-8cc8-4b70-96a9-737019363638-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.997127] env[62368]: DEBUG oslo_concurrency.lockutils [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] Lock "5ed9cee5-8cc8-4b70-96a9-737019363638-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.997348] env[62368]: DEBUG nova.compute.manager [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] No waiting events found dispatching network-vif-plugged-b6d50c17-ef64-4171-8459-4829ffa5d03d {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.997555] env[62368]: WARNING nova.compute.manager [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Received unexpected event network-vif-plugged-b6d50c17-ef64-4171-8459-4829ffa5d03d for instance with vm_state building and task_state spawning. [ 852.997735] env[62368]: DEBUG nova.compute.manager [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Received event network-changed-b6d50c17-ef64-4171-8459-4829ffa5d03d {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.997937] env[62368]: DEBUG nova.compute.manager [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Refreshing instance network info cache due to event network-changed-b6d50c17-ef64-4171-8459-4829ffa5d03d. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.998149] env[62368]: DEBUG oslo_concurrency.lockutils [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] Acquiring lock "refresh_cache-5ed9cee5-8cc8-4b70-96a9-737019363638" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.140026] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198566, 'name': PowerOffVM_Task, 'duration_secs': 0.353055} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.140343] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 853.140673] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 853.141731] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5112a69-9186-44d0-94c0-12a219486c11 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.150251] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 853.150567] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e69bf49-7d2b-4d29-9956-b7e8143712ef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.159094] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.229624] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 853.229891] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 853.230084] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleting the datastore file [datastore1] 58efc1cf-2469-41be-bf99-fc7b8c72113c {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.230481] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5bb8db7d-08ad-41cd-a566-f9bcd681b988 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.237967] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 853.237967] env[62368]: value = "task-1198569" [ 853.237967] env[62368]: _type = "Task" [ 853.237967] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.249655] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.254500] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "refresh_cache-5ed9cee5-8cc8-4b70-96a9-737019363638" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.254890] env[62368]: DEBUG nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Instance network_info: |[{"id": "b6d50c17-ef64-4171-8459-4829ffa5d03d", "address": "fa:16:3e:7d:ed:a7", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6d50c17-ef", "ovs_interfaceid": "b6d50c17-ef64-4171-8459-4829ffa5d03d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.255255] env[62368]: DEBUG oslo_concurrency.lockutils [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] Acquired lock "refresh_cache-5ed9cee5-8cc8-4b70-96a9-737019363638" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.255493] env[62368]: DEBUG nova.network.neutron [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Refreshing network info cache for port b6d50c17-ef64-4171-8459-4829ffa5d03d {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.256927] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:ed:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8f40f5c4-c146-449c-884d-6f884dcf2acf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6d50c17-ef64-4171-8459-4829ffa5d03d', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.265195] env[62368]: DEBUG oslo.service.loopingcall [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.266183] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.267009] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc1d0e69-a2f4-4ad0-a5e4-dd2a0f71b3c2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.283937] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5811f012-4378-4874-bc9d-5c886878ebbe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.905s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.286601] env[62368]: DEBUG nova.scheduler.client.report [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.292357] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.292357] env[62368]: value = "task-1198570" [ 853.292357] env[62368]: _type = "Task" [ 853.292357] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.303467] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198570, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.345472] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198567, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.378169] env[62368]: DEBUG nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.408458] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.408850] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.409042] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.409363] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.409591] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.409824] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.410094] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.410387] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.410592] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.410785] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.411018] env[62368]: DEBUG nova.virt.hardware [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.412108] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5d70fe-8944-4da7-b200-71d8796c9f1a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.421210] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d555ea3-30cc-4687-bbbb-b9f63680a478 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.751494] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198569, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.41964} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.751494] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.751607] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 853.751790] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 853.793977] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.794726] env[62368]: DEBUG nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.797510] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.442s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.797733] env[62368]: DEBUG nova.objects.instance [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lazy-loading 'resources' on Instance uuid 0b861a8b-dcad-46ad-90ef-a46450c5cd51 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.813424] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198570, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.844440] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577927} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.844440] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5/044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 853.844440] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.844891] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b59cf2aa-d279-4e87-ab4c-d42aefb75482 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.852507] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 853.852507] env[62368]: value = "task-1198571" [ 853.852507] env[62368]: _type = "Task" [ 853.852507] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.864074] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.940031] env[62368]: DEBUG oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522984b9-625f-0c2f-cb20-a976ee6424e5/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 853.941279] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2093bfce-bdc4-4c5c-97ec-aefbe1495a20 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.947382] env[62368]: DEBUG oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522984b9-625f-0c2f-cb20-a976ee6424e5/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 853.947577] env[62368]: ERROR oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522984b9-625f-0c2f-cb20-a976ee6424e5/disk-0.vmdk due to incomplete transfer. [ 853.947786] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ad65bff3-17d7-4a99-8d02-c508042bfa9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.954673] env[62368]: DEBUG oslo_vmware.rw_handles [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522984b9-625f-0c2f-cb20-a976ee6424e5/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 853.954868] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Uploaded image 5037b35f-25b2-408c-9f87-c33c3892b67e to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 853.957268] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 853.957511] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ccf37c95-1cff-4fd7-9ebb-c70b7e65f1c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.962656] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 853.962656] env[62368]: value = "task-1198572" [ 853.962656] env[62368]: _type = "Task" [ 853.962656] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.971304] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198572, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.005331] env[62368]: DEBUG nova.network.neutron [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Updated VIF entry in instance network info cache for port b6d50c17-ef64-4171-8459-4829ffa5d03d. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 854.005732] env[62368]: DEBUG nova.network.neutron [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Updating instance_info_cache with network_info: [{"id": "b6d50c17-ef64-4171-8459-4829ffa5d03d", "address": "fa:16:3e:7d:ed:a7", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6d50c17-ef", "ovs_interfaceid": "b6d50c17-ef64-4171-8459-4829ffa5d03d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.025116] env[62368]: DEBUG nova.network.neutron [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Successfully updated port: d6d93849-79aa-4366-a8b1-ce1732983fa4 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.166085] env[62368]: DEBUG nova.network.neutron [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Port 444a47e3-2f56-4895-8a8b-7e5605dbfaf1 binding to destination host cpu-1 is already ACTIVE {{(pid=62368) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 854.166394] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.166554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.166783] env[62368]: DEBUG nova.network.neutron [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.308383] env[62368]: DEBUG nova.compute.utils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.314060] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198570, 'name': CreateVM_Task, 'duration_secs': 0.614644} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.314060] env[62368]: DEBUG nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.314060] env[62368]: DEBUG nova.network.neutron [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 854.315946] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 854.316792] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.317087] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.317518] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.318164] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0056a7d1-767c-4ed1-a7e6-99fcb9ed6110 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.324808] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 854.324808] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524e1756-9945-e85a-809d-72a0e6af965b" [ 854.324808] env[62368]: _type = "Task" [ 854.324808] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.334770] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524e1756-9945-e85a-809d-72a0e6af965b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.363250] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087427} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.363525] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.364302] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4131e8cd-f932-461f-82ed-9a0f6767c329 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.368023] env[62368]: DEBUG nova.policy [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b2339526fa54d94926720a290051dae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da6a595e1dc64441a999ace8ebb22151', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.391151] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5/044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.394045] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82b10c6d-59a0-4ade-b6ce-4f05f7f8d3d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.417066] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 854.417066] env[62368]: value = "task-1198573" [ 854.417066] env[62368]: _type = "Task" [ 854.417066] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.431046] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198573, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.473134] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198572, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.510261] env[62368]: DEBUG oslo_concurrency.lockutils [req-66d83802-7712-400b-9d04-70b6cd7e6832 req-b6eb0d62-c3b1-4d06-aeef-02f16fadbcd8 service nova] Releasing lock "refresh_cache-5ed9cee5-8cc8-4b70-96a9-737019363638" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.528369] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "refresh_cache-f397e102-bfbd-4753-b1a4-4c5e9126b2c6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.528483] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "refresh_cache-f397e102-bfbd-4753-b1a4-4c5e9126b2c6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.528729] env[62368]: DEBUG nova.network.neutron [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.652342] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47865194-c4e0-44d6-a908-ae1a56d99c99 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.655963] env[62368]: DEBUG nova.network.neutron [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Successfully created port: b0fe605d-e7f1-4955-9532-b1e6aa918864 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.664110] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2239aa-1af6-4795-8c6a-d0a11be23841 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.701919] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4212c55a-98b6-475c-9046-60e74ef58555 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.712150] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd6bfe2-91be-49db-bf08-e107d467fcb0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.718400] env[62368]: DEBUG oslo_concurrency.lockutils [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.718645] env[62368]: DEBUG oslo_concurrency.lockutils [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.731135] env[62368]: DEBUG nova.compute.provider_tree [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.733222] env[62368]: DEBUG nova.compute.utils [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.789497] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.789838] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.790037] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.790295] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.790511] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.790705] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.790969] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.791159] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.791338] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.791506] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.791738] env[62368]: DEBUG nova.virt.hardware [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.792787] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9621fcd7-a3d0-4ea0-a70c-09907e5ed5af {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.805535] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6cafd3-9b91-465a-be5b-1b6f421f6316 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.821120] env[62368]: DEBUG nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 854.827416] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:c8:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd3e0156-36ff-4c90-8484-69fed4dfa149', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.831391] env[62368]: DEBUG oslo.service.loopingcall [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.835170] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 854.838612] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9930295-7ec6-4d24-8b2b-c8d932ac5601 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.862770] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524e1756-9945-e85a-809d-72a0e6af965b, 'name': SearchDatastore_Task, 'duration_secs': 0.022069} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.864314] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.864556] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.864845] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.864943] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.865147] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.865390] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.865390] env[62368]: value = "task-1198574" [ 854.865390] env[62368]: _type = "Task" [ 854.865390] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.865593] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9953263f-1bdd-42e8-aac1-9715a71ab77e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.877084] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198574, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.884295] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.884519] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 854.885422] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c8ecb08-60c0-43c9-9892-83cc49394ac0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.891799] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 854.891799] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]525b3802-eac0-6b3d-39af-3ebf42a59dcc" [ 854.891799] env[62368]: _type = "Task" [ 854.891799] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.900909] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525b3802-eac0-6b3d-39af-3ebf42a59dcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.927883] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198573, 'name': ReconfigVM_Task, 'duration_secs': 0.395423} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.928102] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5/044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.928886] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db716aea-a614-4ffa-9c54-10243cc62e71 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.937629] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 854.937629] env[62368]: value = "task-1198575" [ 854.937629] env[62368]: _type = "Task" [ 854.937629] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.947129] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198575, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.973990] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198572, 'name': Destroy_Task, 'duration_secs': 0.848786} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.974304] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Destroyed the VM [ 854.974621] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 854.974893] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-765af9b7-32dd-41c5-97a0-28f88410ba6d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.982548] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 854.982548] env[62368]: value = "task-1198576" [ 854.982548] env[62368]: _type = "Task" [ 854.982548] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.992645] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198576, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.051142] env[62368]: DEBUG nova.network.neutron [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance_info_cache with network_info: [{"id": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "address": "fa:16:3e:98:a0:d2", "network": {"id": "34f156be-cab3-45a2-8b52-74dd4cd1e42d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2a054e975862463bb842479ad19393ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap444a47e3-2f", "ovs_interfaceid": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.067055] env[62368]: DEBUG nova.network.neutron [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.226189] env[62368]: DEBUG nova.network.neutron [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Updating instance_info_cache with network_info: [{"id": "d6d93849-79aa-4366-a8b1-ce1732983fa4", "address": "fa:16:3e:c5:2c:7d", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d93849-79", "ovs_interfaceid": "d6d93849-79aa-4366-a8b1-ce1732983fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.236036] env[62368]: DEBUG nova.scheduler.client.report [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.240116] env[62368]: DEBUG oslo_concurrency.lockutils [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 0.521s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.379814] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198574, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.403698] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525b3802-eac0-6b3d-39af-3ebf42a59dcc, 'name': SearchDatastore_Task, 'duration_secs': 0.046299} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.404512] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88f5130b-583f-4576-ac73-bf034bcba837 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.411139] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 855.411139] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f1ee72-31e9-9855-c693-0db99a4b86a5" [ 855.411139] env[62368]: _type = "Task" [ 855.411139] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.419839] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f1ee72-31e9-9855-c693-0db99a4b86a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.446203] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198575, 'name': Rename_Task, 'duration_secs': 0.301365} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.446874] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 855.447171] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f9b2958-2c80-466a-980e-4956075dcecb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.454371] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 855.454371] env[62368]: value = "task-1198577" [ 855.454371] env[62368]: _type = "Task" [ 855.454371] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.464447] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.494720] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198576, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.515452] env[62368]: DEBUG nova.compute.manager [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Received event network-vif-plugged-d6d93849-79aa-4366-a8b1-ce1732983fa4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.515672] env[62368]: DEBUG oslo_concurrency.lockutils [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] Acquiring lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.515956] env[62368]: DEBUG oslo_concurrency.lockutils [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] Lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.516164] env[62368]: DEBUG oslo_concurrency.lockutils [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] Lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.516339] env[62368]: DEBUG nova.compute.manager [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] No waiting events found dispatching network-vif-plugged-d6d93849-79aa-4366-a8b1-ce1732983fa4 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.516505] env[62368]: WARNING nova.compute.manager [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Received unexpected event network-vif-plugged-d6d93849-79aa-4366-a8b1-ce1732983fa4 for instance with vm_state building and task_state spawning. [ 855.516667] env[62368]: DEBUG nova.compute.manager [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Received event network-changed-d6d93849-79aa-4366-a8b1-ce1732983fa4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.516840] env[62368]: DEBUG nova.compute.manager [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Refreshing instance network info cache due to event network-changed-d6d93849-79aa-4366-a8b1-ce1732983fa4. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 855.516989] env[62368]: DEBUG oslo_concurrency.lockutils [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] Acquiring lock "refresh_cache-f397e102-bfbd-4753-b1a4-4c5e9126b2c6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.553771] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.729205] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "refresh_cache-f397e102-bfbd-4753-b1a4-4c5e9126b2c6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.729563] env[62368]: DEBUG nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Instance network_info: |[{"id": "d6d93849-79aa-4366-a8b1-ce1732983fa4", "address": "fa:16:3e:c5:2c:7d", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d93849-79", "ovs_interfaceid": "d6d93849-79aa-4366-a8b1-ce1732983fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.729898] env[62368]: DEBUG oslo_concurrency.lockutils [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] Acquired lock "refresh_cache-f397e102-bfbd-4753-b1a4-4c5e9126b2c6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.730095] env[62368]: DEBUG nova.network.neutron [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Refreshing network info cache for port d6d93849-79aa-4366-a8b1-ce1732983fa4 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 855.731572] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:2c:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7150f662-0cf1-44f9-ae14-d70f479649b6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6d93849-79aa-4366-a8b1-ce1732983fa4', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.739123] env[62368]: DEBUG oslo.service.loopingcall [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.742354] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 855.743083] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.946s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.745326] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-112f9c4f-09d1-4a25-9f54-37924434e43e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.759952] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 25.551s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.760092] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.760239] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 855.760561] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.019s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.762087] env[62368]: INFO nova.compute.claims [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.765847] env[62368]: INFO nova.scheduler.client.report [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted allocations for instance 0b861a8b-dcad-46ad-90ef-a46450c5cd51 [ 855.767464] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b4ccd6-a659-4404-8664-449a3b7a9d4e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.779992] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8aad38-28c4-4c74-947a-83a39cdcb0ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.784438] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.784438] env[62368]: value = "task-1198578" [ 855.784438] env[62368]: _type = "Task" [ 855.784438] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.802352] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bc855a-c863-4822-b17e-18f7f2ad3c18 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.813764] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198578, 'name': CreateVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.820681] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50dae055-2674-41e8-9849-a18ef1b4e612 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.548224] env[62368]: DEBUG nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.552029] env[62368]: DEBUG nova.compute.manager [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62368) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 856.552242] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.556104] env[62368]: DEBUG oslo_concurrency.lockutils [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.556324] env[62368]: DEBUG oslo_concurrency.lockutils [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.556545] env[62368]: INFO nova.compute.manager [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Attaching volume 92efc3b4-cdd3-4084-8734-f57f630ab2db to /dev/sdb [ 856.558454] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180676MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 856.558580] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.559836] env[62368]: DEBUG nova.network.neutron [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Successfully updated port: b0fe605d-e7f1-4955-9532-b1e6aa918864 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.562780] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ce62245a-340a-4422-8795-99620e005413 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "0b861a8b-dcad-46ad-90ef-a46450c5cd51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.791s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.580745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.580889] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquired lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.581026] env[62368]: DEBUG nova.network.neutron [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.585091] env[62368]: DEBUG oslo_concurrency.lockutils [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.585268] env[62368]: DEBUG oslo_concurrency.lockutils [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.599937] env[62368]: DEBUG oslo_vmware.api [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198576, 'name': RemoveSnapshot_Task, 'duration_secs': 1.155499} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.600139] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198574, 'name': CreateVM_Task, 'duration_secs': 0.638407} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.600595] env[62368]: DEBUG oslo_vmware.api [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198577, 'name': PowerOnVM_Task, 'duration_secs': 0.709561} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.608126] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 856.608370] env[62368]: INFO nova.compute.manager [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Took 17.11 seconds to snapshot the instance on the hypervisor. [ 856.610572] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 856.611082] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 856.611293] env[62368]: INFO nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Took 8.26 seconds to spawn the instance on the hypervisor. [ 856.611474] env[62368]: DEBUG nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.611715] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198578, 'name': CreateVM_Task, 'duration_secs': 0.442794} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.611943] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f1ee72-31e9-9855-c693-0db99a4b86a5, 'name': SearchDatastore_Task, 'duration_secs': 0.011992} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.614839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.614999] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.615337] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.616091] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365e96ed-8713-404f-ba6c-faa6bfec6c1a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.618493] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 856.618763] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.619000] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 5ed9cee5-8cc8-4b70-96a9-737019363638/5ed9cee5-8cc8-4b70-96a9-737019363638.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 856.620294] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-526f354e-151c-4336-9005-b7b1ef805275 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.625730] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.625730] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf07f792-2d33-4037-a317-1f70ec380e36 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.625730] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c14767-7a10-4239-99d9-6f667efc180d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.629862] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.630094] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.630253] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.630446] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.630594] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.631084] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.631084] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.631191] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.631288] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.631429] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.631606] env[62368]: DEBUG nova.virt.hardware [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.632960] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b289f85-ee74-43e4-936b-11d6fe75dcef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.644020] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 856.644020] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c8b72b-caa0-d8ca-9bbe-4988388f929b" [ 856.644020] env[62368]: _type = "Task" [ 856.644020] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.649645] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230ab3d6-9602-4430-93df-706d0f0bc9a4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.652697] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 856.652697] env[62368]: value = "task-1198579" [ 856.652697] env[62368]: _type = "Task" [ 856.652697] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.656940] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52530a07-37b6-446e-bdae-204b91adb222 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.668635] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c8b72b-caa0-d8ca-9bbe-4988388f929b, 'name': SearchDatastore_Task, 'duration_secs': 0.013489} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.669229] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.669311] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.670104] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.670104] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.670104] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.670591] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.670938] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.671438] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c462f22e-6c42-49c5-8a0e-d8bbe2a6b8c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.687564] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4bf27dd-6b7c-425a-9beb-a8f7cda37ccf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.690203] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198579, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.690648] env[62368]: DEBUG nova.virt.block_device [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Updating existing volume attachment record: a6aef9e4-0d28-4876-a134-c06f8c60bd8e {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 856.699126] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 856.699126] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524cfd57-343e-9310-bea4-450d4736bcf5" [ 856.699126] env[62368]: _type = "Task" [ 856.699126] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.703649] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.703876] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 856.704990] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d283f0f-aae4-45de-b7b1-599a31400962 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.710867] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524cfd57-343e-9310-bea4-450d4736bcf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.714504] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 856.714504] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ab13a6-969c-3167-ceab-88a4ddcee44e" [ 856.714504] env[62368]: _type = "Task" [ 856.714504] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.721465] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ab13a6-969c-3167-ceab-88a4ddcee44e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.806064] env[62368]: DEBUG nova.network.neutron [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Updated VIF entry in instance network info cache for port d6d93849-79aa-4366-a8b1-ce1732983fa4. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 856.806489] env[62368]: DEBUG nova.network.neutron [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Updating instance_info_cache with network_info: [{"id": "d6d93849-79aa-4366-a8b1-ce1732983fa4", "address": "fa:16:3e:c5:2c:7d", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d93849-79", "ovs_interfaceid": "d6d93849-79aa-4366-a8b1-ce1732983fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.090022] env[62368]: DEBUG nova.compute.utils [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.150259] env[62368]: DEBUG nova.network.neutron [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 857.159834] env[62368]: INFO nova.compute.manager [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Took 43.74 seconds to build instance. [ 857.169394] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198579, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.178721] env[62368]: DEBUG nova.compute.manager [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Found 3 images (rotation: 2) {{(pid=62368) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 857.178927] env[62368]: DEBUG nova.compute.manager [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Rotating out 1 backups {{(pid=62368) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 857.179356] env[62368]: DEBUG nova.compute.manager [None req-f44b63d7-2144-4170-9b54-a4e8c07cf406 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleting image 3b178833-ecf1-4a1f-8ffe-ee20324be047 {{(pid=62368) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 857.214256] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524cfd57-343e-9310-bea4-450d4736bcf5, 'name': SearchDatastore_Task, 'duration_secs': 0.019012} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.214886] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.215199] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.215455] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.229583] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ab13a6-969c-3167-ceab-88a4ddcee44e, 'name': SearchDatastore_Task, 'duration_secs': 0.009918} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.230238] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-541a4641-dd98-494a-a7c4-a140c1e46712 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.239841] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 857.239841] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ac74c1-aa61-8306-c2ef-7d9f71011a19" [ 857.239841] env[62368]: _type = "Task" [ 857.239841] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.249577] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ac74c1-aa61-8306-c2ef-7d9f71011a19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.309103] env[62368]: DEBUG oslo_concurrency.lockutils [req-edaf1437-dcb6-43d0-9d18-ba82f0bf01fe req-5221b099-d7fb-4aff-b9bc-7d8fd152be0d service nova] Releasing lock "refresh_cache-f397e102-bfbd-4753-b1a4-4c5e9126b2c6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.412060] env[62368]: DEBUG nova.network.neutron [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Updating instance_info_cache with network_info: [{"id": "b0fe605d-e7f1-4955-9532-b1e6aa918864", "address": "fa:16:3e:9e:4a:a2", "network": {"id": "2038d073-bf70-4f85-adfa-cf1d464d908a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-561693200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da6a595e1dc64441a999ace8ebb22151", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0fe605d-e7", "ovs_interfaceid": "b0fe605d-e7f1-4955-9532-b1e6aa918864", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.430224] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5443188-b7c1-48f6-95ea-cf8caeb108ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.439250] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf479cd-b05c-42bd-89df-399de995ba9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.473534] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3f6d63-dab9-4ad5-bec0-566d982612c9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.482333] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0e8c98-4d1a-40d9-b34a-a48fbd019b0c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.496464] env[62368]: DEBUG nova.compute.provider_tree [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.595293] env[62368]: DEBUG oslo_concurrency.lockutils [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.661909] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2c67380f-8d71-4bec-b778-2fbca91ec62e tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.878s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.666525] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198579, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.773624} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.666787] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 5ed9cee5-8cc8-4b70-96a9-737019363638/5ed9cee5-8cc8-4b70-96a9-737019363638.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 857.666996] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.667264] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41707071-e38b-42c7-911b-2747ec795641 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.673923] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 857.673923] env[62368]: value = "task-1198583" [ 857.673923] env[62368]: _type = "Task" [ 857.673923] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.683393] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198583, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.752548] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ac74c1-aa61-8306-c2ef-7d9f71011a19, 'name': SearchDatastore_Task, 'duration_secs': 0.052931} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.752843] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.753125] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 58efc1cf-2469-41be-bf99-fc7b8c72113c/58efc1cf-2469-41be-bf99-fc7b8c72113c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 857.753427] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.753620] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.753824] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15925aea-bdca-4626-88ad-d09118d95018 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.755750] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-671bd5e0-99f1-446a-90b7-92e2c26c32a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.763680] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 857.763680] env[62368]: value = "task-1198584" [ 857.763680] env[62368]: _type = "Task" [ 857.763680] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.768747] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.768747] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 857.768941] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2506a3f9-f432-4658-ad93-a496496c86b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.773911] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198584, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.777835] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 857.777835] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ad0764-c7f5-ef00-53dd-90292e7c990f" [ 857.777835] env[62368]: _type = "Task" [ 857.777835] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.785464] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ad0764-c7f5-ef00-53dd-90292e7c990f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.915283] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Releasing lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.915751] env[62368]: DEBUG nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Instance network_info: |[{"id": "b0fe605d-e7f1-4955-9532-b1e6aa918864", "address": "fa:16:3e:9e:4a:a2", "network": {"id": "2038d073-bf70-4f85-adfa-cf1d464d908a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-561693200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da6a595e1dc64441a999ace8ebb22151", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0fe605d-e7", "ovs_interfaceid": "b0fe605d-e7f1-4955-9532-b1e6aa918864", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.916121] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:4a:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc6e6fe1-c4f5-4389-a49f-0978060eebb4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0fe605d-e7f1-4955-9532-b1e6aa918864', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.923635] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Creating folder: Project (da6a595e1dc64441a999ace8ebb22151). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 857.923938] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94968b9e-8a94-4f14-8b1c-ff8e0f6f03e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.938326] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Created folder: Project (da6a595e1dc64441a999ace8ebb22151) in parent group-v259706. [ 857.938607] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Creating folder: Instances. Parent ref: group-v259830. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 857.938955] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87686d86-8ba4-4da5-9e8a-82260264389e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.951758] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Created folder: Instances in parent group-v259830. [ 857.952058] env[62368]: DEBUG oslo.service.loopingcall [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.952282] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 857.952627] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34a994d5-c47d-4da2-8173-74d122f79ba8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.973789] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.973789] env[62368]: value = "task-1198587" [ 857.973789] env[62368]: _type = "Task" [ 857.973789] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.985703] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198587, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.999894] env[62368]: DEBUG nova.scheduler.client.report [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.008022] env[62368]: DEBUG nova.compute.manager [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Received event network-vif-plugged-b0fe605d-e7f1-4955-9532-b1e6aa918864 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.008150] env[62368]: DEBUG oslo_concurrency.lockutils [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] Acquiring lock "0cb0097b-2908-4107-bb57-0241915ccff7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.008370] env[62368]: DEBUG oslo_concurrency.lockutils [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] Lock "0cb0097b-2908-4107-bb57-0241915ccff7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.008753] env[62368]: DEBUG oslo_concurrency.lockutils [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] Lock "0cb0097b-2908-4107-bb57-0241915ccff7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.009200] env[62368]: DEBUG nova.compute.manager [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] No waiting events found dispatching network-vif-plugged-b0fe605d-e7f1-4955-9532-b1e6aa918864 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.009200] env[62368]: WARNING nova.compute.manager [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Received unexpected event network-vif-plugged-b0fe605d-e7f1-4955-9532-b1e6aa918864 for instance with vm_state building and task_state spawning. [ 858.009308] env[62368]: DEBUG nova.compute.manager [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Received event network-changed-b0fe605d-e7f1-4955-9532-b1e6aa918864 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.009837] env[62368]: DEBUG nova.compute.manager [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Refreshing instance network info cache due to event network-changed-b0fe605d-e7f1-4955-9532-b1e6aa918864. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.009837] env[62368]: DEBUG oslo_concurrency.lockutils [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] Acquiring lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.009837] env[62368]: DEBUG oslo_concurrency.lockutils [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] Acquired lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.010064] env[62368]: DEBUG nova.network.neutron [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Refreshing network info cache for port b0fe605d-e7f1-4955-9532-b1e6aa918864 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.102242] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "417f000f-cf23-404d-877c-45990d1a7c77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.102578] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "417f000f-cf23-404d-877c-45990d1a7c77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.102817] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "417f000f-cf23-404d-877c-45990d1a7c77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.103053] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "417f000f-cf23-404d-877c-45990d1a7c77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.103294] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "417f000f-cf23-404d-877c-45990d1a7c77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.105896] env[62368]: INFO nova.compute.manager [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Terminating instance [ 858.108107] env[62368]: DEBUG nova.compute.manager [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 858.108212] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 858.109407] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0f71ff-537b-4d87-a2b6-1fa539c0fdc4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.119279] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 858.119557] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23bd6b98-48c1-472d-b663-52554bfebc35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.130260] env[62368]: DEBUG oslo_vmware.api [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 858.130260] env[62368]: value = "task-1198588" [ 858.130260] env[62368]: _type = "Task" [ 858.130260] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.141922] env[62368]: DEBUG oslo_vmware.api [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198588, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.190071] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198583, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071876} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.190622] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.191865] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e81597-3bcb-4de3-b722-298974518e04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.230460] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 5ed9cee5-8cc8-4b70-96a9-737019363638/5ed9cee5-8cc8-4b70-96a9-737019363638.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.230829] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6118dca6-0654-4cc0-b944-2dd7a33886c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.254197] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 858.254197] env[62368]: value = "task-1198589" [ 858.254197] env[62368]: _type = "Task" [ 858.254197] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.264073] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198589, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.275202] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198584, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.290727] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ad0764-c7f5-ef00-53dd-90292e7c990f, 'name': SearchDatastore_Task, 'duration_secs': 0.013093} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.291843] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c498b745-434c-4558-8d3e-1ad3be5c5cf9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.299141] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 858.299141] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ba6716-dfc4-9ead-0427-36b15d62f123" [ 858.299141] env[62368]: _type = "Task" [ 858.299141] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.310503] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ba6716-dfc4-9ead-0427-36b15d62f123, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.484706] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198587, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.505881] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.745s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.506565] env[62368]: DEBUG nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 858.509663] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.366s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.510917] env[62368]: INFO nova.compute.claims [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.563701] env[62368]: DEBUG oslo_concurrency.lockutils [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.564011] env[62368]: DEBUG oslo_concurrency.lockutils [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.564247] env[62368]: INFO nova.compute.manager [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Attaching volume 11c38512-c804-4553-9c51-d6a79caa8f45 to /dev/sdb [ 858.602182] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6873a7c-3fc2-4b3e-9a55-67ea021765bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.609911] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115e08e4-2685-40c6-b15a-4e5c3ecfda65 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.631315] env[62368]: DEBUG nova.virt.block_device [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updating existing volume attachment record: 53aedaf8-66a0-471d-880e-82a99f743334 {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 858.641571] env[62368]: DEBUG oslo_vmware.api [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198588, 'name': PowerOffVM_Task, 'duration_secs': 0.363472} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.641571] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 858.641678] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 858.642184] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9db35687-1d10-44ff-ac5c-5ec1a7205f84 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.724202] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 858.724202] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 858.724202] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleting the datastore file [datastore2] 417f000f-cf23-404d-877c-45990d1a7c77 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.730061] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec820e23-f2c2-49d5-aeed-7de5e4ab3767 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.738034] env[62368]: DEBUG oslo_vmware.api [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 858.738034] env[62368]: value = "task-1198591" [ 858.738034] env[62368]: _type = "Task" [ 858.738034] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.750150] env[62368]: DEBUG oslo_vmware.api [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.764748] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198589, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.771015] env[62368]: DEBUG nova.network.neutron [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Updated VIF entry in instance network info cache for port b0fe605d-e7f1-4955-9532-b1e6aa918864. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 858.771401] env[62368]: DEBUG nova.network.neutron [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Updating instance_info_cache with network_info: [{"id": "b0fe605d-e7f1-4955-9532-b1e6aa918864", "address": "fa:16:3e:9e:4a:a2", "network": {"id": "2038d073-bf70-4f85-adfa-cf1d464d908a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-561693200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da6a595e1dc64441a999ace8ebb22151", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0fe605d-e7", "ovs_interfaceid": "b0fe605d-e7f1-4955-9532-b1e6aa918864", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.782051] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198584, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.684808} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.782051] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 58efc1cf-2469-41be-bf99-fc7b8c72113c/58efc1cf-2469-41be-bf99-fc7b8c72113c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 858.782051] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.782051] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-667f45cb-42cb-4d62-a6b7-91fde420649d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.791983] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 858.791983] env[62368]: value = "task-1198593" [ 858.791983] env[62368]: _type = "Task" [ 858.791983] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.802437] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198593, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.813223] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ba6716-dfc4-9ead-0427-36b15d62f123, 'name': SearchDatastore_Task, 'duration_secs': 0.058374} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.813305] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.813638] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] f397e102-bfbd-4753-b1a4-4c5e9126b2c6/f397e102-bfbd-4753-b1a4-4c5e9126b2c6.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 858.813970] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fdb0c25d-0f15-4145-9100-0f00993bb319 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.821834] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 858.821834] env[62368]: value = "task-1198594" [ 858.821834] env[62368]: _type = "Task" [ 858.821834] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.833136] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198594, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.986861] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198587, 'name': CreateVM_Task, 'duration_secs': 0.734089} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.987102] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 858.987732] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.987947] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.988493] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.988789] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f234e91-8199-4252-b4fc-3fab344275fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.994505] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 858.994505] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f836e4-b843-6a9e-f1f6-a648f9d31726" [ 858.994505] env[62368]: _type = "Task" [ 858.994505] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.003674] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f836e4-b843-6a9e-f1f6-a648f9d31726, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.015533] env[62368]: DEBUG nova.compute.utils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.017208] env[62368]: DEBUG nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.017385] env[62368]: DEBUG nova.network.neutron [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 859.088988] env[62368]: DEBUG nova.policy [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32da75d294d548eb8c5f75ae18e79c0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3940819fc14a4c628aacd8820efe084e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.248375] env[62368]: DEBUG oslo_vmware.api [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198591, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288475} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.248536] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.248728] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 859.248909] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 859.249099] env[62368]: INFO nova.compute.manager [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Took 1.14 seconds to destroy the instance on the hypervisor. [ 859.249376] env[62368]: DEBUG oslo.service.loopingcall [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.249948] env[62368]: DEBUG nova.compute.manager [-] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 859.250088] env[62368]: DEBUG nova.network.neutron [-] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 859.265703] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198589, 'name': ReconfigVM_Task, 'duration_secs': 0.69374} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.266080] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 5ed9cee5-8cc8-4b70-96a9-737019363638/5ed9cee5-8cc8-4b70-96a9-737019363638.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.266967] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f946c20-ff2e-4cae-8680-8fa08dd20cdf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.276787] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 859.276787] env[62368]: value = "task-1198598" [ 859.276787] env[62368]: _type = "Task" [ 859.276787] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.281157] env[62368]: DEBUG oslo_concurrency.lockutils [req-2d3d7cba-8931-4a5d-8084-14b716a70ac5 req-a39426d6-a435-4a56-9cad-f60cf8ba2ea2 service nova] Releasing lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.289177] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198598, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.304552] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198593, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072661} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.307218] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.311884] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882685ec-01ee-4013-be36-b13353058769 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.315884] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "55eecf3d-501c-490d-bddd-0211fd082841" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.316236] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "55eecf3d-501c-490d-bddd-0211fd082841" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.342042] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 58efc1cf-2469-41be-bf99-fc7b8c72113c/58efc1cf-2469-41be-bf99-fc7b8c72113c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.349658] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6767229d-7d90-4d27-9bbe-c0fd126083f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.372782] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198594, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.374535] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 859.374535] env[62368]: value = "task-1198599" [ 859.374535] env[62368]: _type = "Task" [ 859.374535] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.386046] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198599, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.505445] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f836e4-b843-6a9e-f1f6-a648f9d31726, 'name': SearchDatastore_Task, 'duration_secs': 0.01612} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.505835] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.507017] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.507017] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.507017] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.507017] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.507017] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd7e17f1-72a2-4bd2-8a65-3548e80a6dc2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.517594] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.517804] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 859.519071] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c05ed3dc-3e40-47a3-ae76-7f8dcb1eeade {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.524024] env[62368]: DEBUG nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 859.535401] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 859.535401] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52756888-5981-72c6-2a3e-48ceb685d247" [ 859.535401] env[62368]: _type = "Task" [ 859.535401] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.546505] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52756888-5981-72c6-2a3e-48ceb685d247, 'name': SearchDatastore_Task, 'duration_secs': 0.018159} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.547452] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2111edbf-1851-4ec0-91c9-91cbce0ef5c3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.555690] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 859.555690] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52d8c94a-13ad-e799-7e5c-6fb9bc1aa08e" [ 859.555690] env[62368]: _type = "Task" [ 859.555690] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.569852] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d8c94a-13ad-e799-7e5c-6fb9bc1aa08e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.606563] env[62368]: DEBUG nova.network.neutron [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Successfully created port: 3b605458-f2bc-4c5e-8d9a-44fd80ab88bf {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.707024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquiring lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.707024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.707024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquiring lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.707024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.707024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.712145] env[62368]: INFO nova.compute.manager [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Terminating instance [ 859.722783] env[62368]: DEBUG nova.compute.manager [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 859.723229] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 859.725104] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115f898b-b5f1-4b81-8ca2-c626e2e3ca34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.735509] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 859.735844] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-932ca3a1-2e77-42f9-a680-5065ef118d69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.744129] env[62368]: DEBUG oslo_vmware.api [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 859.744129] env[62368]: value = "task-1198600" [ 859.744129] env[62368]: _type = "Task" [ 859.744129] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.756074] env[62368]: DEBUG oslo_vmware.api [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.791085] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198598, 'name': Rename_Task, 'duration_secs': 0.39319} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.791085] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 859.791085] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d86d668d-23dc-472c-bb01-d21449ed6bcf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.797735] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 859.797735] env[62368]: value = "task-1198601" [ 859.797735] env[62368]: _type = "Task" [ 859.797735] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.806418] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198601, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.818802] env[62368]: DEBUG nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 859.841490] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198594, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.650231} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.841779] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] f397e102-bfbd-4753-b1a4-4c5e9126b2c6/f397e102-bfbd-4753-b1a4-4c5e9126b2c6.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 859.841999] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.843599] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f82b480-629e-4051-8405-c02a9d654d48 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.852902] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 859.852902] env[62368]: value = "task-1198602" [ 859.852902] env[62368]: _type = "Task" [ 859.852902] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.869738] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198602, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.891208] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198599, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.975493] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3cbf45-ede9-426f-bbf0-ef4181a58901 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.989712] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7bae103-f494-45f5-9b04-261cecf8e9d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.026549] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e73c828-0cab-409a-afc9-996c869eb41f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.034729] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcd1008-1625-4249-915b-bb1757a91356 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.055354] env[62368]: DEBUG nova.compute.provider_tree [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.066989] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d8c94a-13ad-e799-7e5c-6fb9bc1aa08e, 'name': SearchDatastore_Task, 'duration_secs': 0.027197} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.068759] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.068759] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 0cb0097b-2908-4107-bb57-0241915ccff7/0cb0097b-2908-4107-bb57-0241915ccff7.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 860.068759] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2a2320c-b9cb-49a0-8476-357e5073d7ef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.077217] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 860.077217] env[62368]: value = "task-1198603" [ 860.077217] env[62368]: _type = "Task" [ 860.077217] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.086906] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198603, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.124043] env[62368]: DEBUG nova.compute.manager [req-8aeae08c-a9a6-472c-bf91-c08be7f6cc53 req-04fba921-03ab-4914-9662-70c4a1f063f0 service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Received event network-vif-deleted-6e4af274-eb1c-4a81-b1f6-dc415b7be3d5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.124343] env[62368]: INFO nova.compute.manager [req-8aeae08c-a9a6-472c-bf91-c08be7f6cc53 req-04fba921-03ab-4914-9662-70c4a1f063f0 service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Neutron deleted interface 6e4af274-eb1c-4a81-b1f6-dc415b7be3d5; detaching it from the instance and deleting it from the info cache [ 860.124711] env[62368]: DEBUG nova.network.neutron [req-8aeae08c-a9a6-472c-bf91-c08be7f6cc53 req-04fba921-03ab-4914-9662-70c4a1f063f0 service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.254803] env[62368]: DEBUG oslo_vmware.api [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198600, 'name': PowerOffVM_Task, 'duration_secs': 0.456419} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.255189] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 860.255436] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 860.255757] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d38ecf9-6ac5-4354-a12c-f3eb64f35ed5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.309671] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198601, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.347237] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.347823] env[62368]: DEBUG nova.network.neutron [-] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.350639] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 860.350859] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 860.351056] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Deleting the datastore file [datastore2] 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.351526] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de370e56-3d75-4a5e-822a-ba1a241b17e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.362849] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198602, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076865} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.364065] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.364421] env[62368]: DEBUG oslo_vmware.api [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for the task: (returnval){ [ 860.364421] env[62368]: value = "task-1198605" [ 860.364421] env[62368]: _type = "Task" [ 860.364421] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.365154] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f87eef4-417e-4b1e-a55c-1982228ce194 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.393046] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] f397e102-bfbd-4753-b1a4-4c5e9126b2c6/f397e102-bfbd-4753-b1a4-4c5e9126b2c6.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.400693] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf4f0028-ed11-4ccd-b89c-b200bf01e1ab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.416046] env[62368]: DEBUG oslo_vmware.api [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198605, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.423297] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198599, 'name': ReconfigVM_Task, 'duration_secs': 0.641981} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.425269] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 58efc1cf-2469-41be-bf99-fc7b8c72113c/58efc1cf-2469-41be-bf99-fc7b8c72113c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.426052] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 860.426052] env[62368]: value = "task-1198606" [ 860.426052] env[62368]: _type = "Task" [ 860.426052] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.426430] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb635812-acfd-4ab1-9230-0a487b3bfffc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.438395] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198606, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.440052] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 860.440052] env[62368]: value = "task-1198607" [ 860.440052] env[62368]: _type = "Task" [ 860.440052] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.453975] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198607, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.546017] env[62368]: DEBUG nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 860.562551] env[62368]: DEBUG nova.scheduler.client.report [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.577135] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.577583] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.577770] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.578067] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.578223] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.578444] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.578812] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.579099] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.579255] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.579484] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.579848] env[62368]: DEBUG nova.virt.hardware [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.581044] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d4e61b-77f7-40c1-981e-dd22f07ab926 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.598575] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a9a619-d245-4f2b-b11f-49fb3c337bfe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.604068] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198603, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.627364] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-838cdcf6-a16c-4ec0-a4f4-dfe5bc5cbf2e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.639848] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50b5024-2221-4e33-97d9-ad26843ced33 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.676549] env[62368]: DEBUG nova.compute.manager [req-8aeae08c-a9a6-472c-bf91-c08be7f6cc53 req-04fba921-03ab-4914-9662-70c4a1f063f0 service nova] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Detach interface failed, port_id=6e4af274-eb1c-4a81-b1f6-dc415b7be3d5, reason: Instance 417f000f-cf23-404d-877c-45990d1a7c77 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 860.810708] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198601, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.836872] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.836872] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.836872] env[62368]: DEBUG nova.compute.manager [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.837312] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd17ae33-8fc5-4d3b-9d43-c5c4811d7762 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.845257] env[62368]: DEBUG nova.compute.manager [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62368) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 860.845931] env[62368]: DEBUG nova.objects.instance [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'flavor' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.852209] env[62368]: INFO nova.compute.manager [-] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Took 1.60 seconds to deallocate network for instance. [ 860.877870] env[62368]: DEBUG oslo_vmware.api [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Task: {'id': task-1198605, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.512902} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.879031] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.879031] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 860.879031] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 860.879031] env[62368]: INFO nova.compute.manager [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 860.879340] env[62368]: DEBUG oslo.service.loopingcall [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.879431] env[62368]: DEBUG nova.compute.manager [-] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.879528] env[62368]: DEBUG nova.network.neutron [-] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 860.945357] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198606, 'name': ReconfigVM_Task, 'duration_secs': 0.376962} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.945357] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Reconfigured VM instance instance-0000004c to attach disk [datastore2] f397e102-bfbd-4753-b1a4-4c5e9126b2c6/f397e102-bfbd-4753-b1a4-4c5e9126b2c6.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.947127] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3ef19e8-68e9-42d8-8f1b-46329329dce6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.953156] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198607, 'name': Rename_Task, 'duration_secs': 0.416588} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.954627] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 860.954968] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 860.954968] env[62368]: value = "task-1198608" [ 860.954968] env[62368]: _type = "Task" [ 860.954968] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.955186] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8903ee1-acaf-4198-a068-1d8d8c03cb97 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.966079] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198608, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.967575] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 860.967575] env[62368]: value = "task-1198610" [ 860.967575] env[62368]: _type = "Task" [ 860.967575] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.983547] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198610, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.071519] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.072301] env[62368]: DEBUG nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.075402] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.106s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.075672] env[62368]: DEBUG nova.objects.instance [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lazy-loading 'resources' on Instance uuid d865d433-2341-4f6c-b840-609530ae2e51 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.092401] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198603, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.692144} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.092401] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 0cb0097b-2908-4107-bb57-0241915ccff7/0cb0097b-2908-4107-bb57-0241915ccff7.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.092401] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.092401] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e7f6ce3-8cd8-4de3-a2d6-db5a0917ea59 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.099696] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 861.099696] env[62368]: value = "task-1198611" [ 861.099696] env[62368]: _type = "Task" [ 861.099696] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.109559] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198611, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.251734] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 861.252068] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259829', 'volume_id': '92efc3b4-cdd3-4084-8734-f57f630ab2db', 'name': 'volume-92efc3b4-cdd3-4084-8734-f57f630ab2db', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5554732b-34a1-41bd-9d0d-9bd39d62bec3', 'attached_at': '', 'detached_at': '', 'volume_id': '92efc3b4-cdd3-4084-8734-f57f630ab2db', 'serial': '92efc3b4-cdd3-4084-8734-f57f630ab2db'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 861.253585] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c02d91c1-1d94-4ae1-a77c-78cbddbc97e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.279427] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e7ad40-00f1-42cd-a38f-513755685cfb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.310466] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] volume-92efc3b4-cdd3-4084-8734-f57f630ab2db/volume-92efc3b4-cdd3-4084-8734-f57f630ab2db.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.316200] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6407585f-f224-48bb-913e-0ad8680e7dc7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.335301] env[62368]: DEBUG oslo_vmware.api [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198601, 'name': PowerOnVM_Task, 'duration_secs': 1.225368} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.336993] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.337286] env[62368]: INFO nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Took 10.49 seconds to spawn the instance on the hypervisor. [ 861.337528] env[62368]: DEBUG nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.337910] env[62368]: DEBUG oslo_vmware.api [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 861.337910] env[62368]: value = "task-1198612" [ 861.337910] env[62368]: _type = "Task" [ 861.337910] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.338682] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f150f29-ded7-401b-88b9-abd1e8cbd2d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.355131] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 861.356315] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c129de93-9298-40ab-b31e-d5ab85cec9a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.358205] env[62368]: DEBUG oslo_vmware.api [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198612, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.360175] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.365510] env[62368]: DEBUG oslo_vmware.api [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 861.365510] env[62368]: value = "task-1198613" [ 861.365510] env[62368]: _type = "Task" [ 861.365510] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.382022] env[62368]: DEBUG oslo_vmware.api [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198613, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.475962] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198608, 'name': Rename_Task, 'duration_secs': 0.179734} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.476743] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 861.477074] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a26aff5-edf5-4663-aa70-936e9bd37ac0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.482512] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198610, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.486255] env[62368]: DEBUG nova.compute.manager [req-144552ea-d54c-4848-9982-22a2932f61cb req-c36a9c40-1d0d-486a-825d-14463e9ad536 service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Received event network-vif-plugged-3b605458-f2bc-4c5e-8d9a-44fd80ab88bf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.486417] env[62368]: DEBUG oslo_concurrency.lockutils [req-144552ea-d54c-4848-9982-22a2932f61cb req-c36a9c40-1d0d-486a-825d-14463e9ad536 service nova] Acquiring lock "b87094da-6258-469e-ab37-5557955ad3a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.486668] env[62368]: DEBUG oslo_concurrency.lockutils [req-144552ea-d54c-4848-9982-22a2932f61cb req-c36a9c40-1d0d-486a-825d-14463e9ad536 service nova] Lock "b87094da-6258-469e-ab37-5557955ad3a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.486886] env[62368]: DEBUG oslo_concurrency.lockutils [req-144552ea-d54c-4848-9982-22a2932f61cb req-c36a9c40-1d0d-486a-825d-14463e9ad536 service nova] Lock "b87094da-6258-469e-ab37-5557955ad3a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.487099] env[62368]: DEBUG nova.compute.manager [req-144552ea-d54c-4848-9982-22a2932f61cb req-c36a9c40-1d0d-486a-825d-14463e9ad536 service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] No waiting events found dispatching network-vif-plugged-3b605458-f2bc-4c5e-8d9a-44fd80ab88bf {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.487357] env[62368]: WARNING nova.compute.manager [req-144552ea-d54c-4848-9982-22a2932f61cb req-c36a9c40-1d0d-486a-825d-14463e9ad536 service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Received unexpected event network-vif-plugged-3b605458-f2bc-4c5e-8d9a-44fd80ab88bf for instance with vm_state building and task_state spawning. [ 861.488062] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 861.488062] env[62368]: value = "task-1198614" [ 861.488062] env[62368]: _type = "Task" [ 861.488062] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.497986] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198614, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.579165] env[62368]: DEBUG nova.compute.utils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.580727] env[62368]: DEBUG nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 861.581210] env[62368]: DEBUG nova.network.neutron [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 861.590819] env[62368]: DEBUG nova.network.neutron [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Successfully updated port: 3b605458-f2bc-4c5e-8d9a-44fd80ab88bf {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.611979] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198611, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.215435} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.613549] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.614447] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dff591d-a733-4cd1-8c36-60bb55563c0d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.640032] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 0cb0097b-2908-4107-bb57-0241915ccff7/0cb0097b-2908-4107-bb57-0241915ccff7.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.643593] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91838b58-e665-4d46-ba20-0474dcbf8549 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.662096] env[62368]: DEBUG nova.policy [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32da75d294d548eb8c5f75ae18e79c0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3940819fc14a4c628aacd8820efe084e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.671751] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 861.671751] env[62368]: value = "task-1198615" [ 861.671751] env[62368]: _type = "Task" [ 861.671751] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.684925] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198615, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.720146] env[62368]: DEBUG nova.network.neutron [-] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.857327] env[62368]: DEBUG oslo_vmware.api [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.870322] env[62368]: INFO nova.compute.manager [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Took 44.34 seconds to build instance. [ 861.877785] env[62368]: DEBUG oslo_vmware.api [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198613, 'name': PowerOffVM_Task, 'duration_secs': 0.227443} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.880566] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 861.880806] env[62368]: DEBUG nova.compute.manager [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.881898] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2614930a-cb44-4642-a38e-009279d256ac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.986018] env[62368]: DEBUG oslo_vmware.api [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198610, 'name': PowerOnVM_Task, 'duration_secs': 0.909362} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.987228] env[62368]: DEBUG nova.network.neutron [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Successfully created port: cb4b1e4c-15e5-4641-b434-e05e31283e5c {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.992969] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.993306] env[62368]: DEBUG nova.compute.manager [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.994661] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3946bb30-7daa-4967-91d5-92f37a7465d9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.008940] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198614, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.029676] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d13d7f-d994-4727-8bae-bba36d270101 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.038844] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a77fe03f-ea1f-45fc-83de-8bb1854a378f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.078911] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2109ad5b-2c62-4803-8116-d01c6880b0ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.088291] env[62368]: DEBUG nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.093836] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.094149] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.094244] env[62368]: DEBUG nova.network.neutron [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.096628] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60ca3db-601f-46b0-969a-26914fe3e004 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.116423] env[62368]: DEBUG nova.compute.provider_tree [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.183140] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198615, 'name': ReconfigVM_Task, 'duration_secs': 0.505134} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.183140] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 0cb0097b-2908-4107-bb57-0241915ccff7/0cb0097b-2908-4107-bb57-0241915ccff7.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.183499] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2d3612b-0f6e-415c-a2c2-698ed80b63e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.190981] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 862.190981] env[62368]: value = "task-1198616" [ 862.190981] env[62368]: _type = "Task" [ 862.190981] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.204989] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198616, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.222818] env[62368]: INFO nova.compute.manager [-] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Took 1.34 seconds to deallocate network for instance. [ 862.254621] env[62368]: DEBUG nova.compute.manager [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Received event network-vif-deleted-932cae7b-da6e-406b-8d5a-a5e049d2cffd {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.254875] env[62368]: DEBUG nova.compute.manager [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Received event network-changed-3b605458-f2bc-4c5e-8d9a-44fd80ab88bf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.255175] env[62368]: DEBUG nova.compute.manager [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Refreshing instance network info cache due to event network-changed-3b605458-f2bc-4c5e-8d9a-44fd80ab88bf. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.255509] env[62368]: DEBUG oslo_concurrency.lockutils [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] Acquiring lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.357013] env[62368]: DEBUG oslo_vmware.api [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198612, 'name': ReconfigVM_Task, 'duration_secs': 0.76485} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.357330] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Reconfigured VM instance instance-00000049 to attach disk [datastore2] volume-92efc3b4-cdd3-4084-8734-f57f630ab2db/volume-92efc3b4-cdd3-4084-8734-f57f630ab2db.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.362571] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c2aed5e-1d32-45a5-84b0-ca75b691abd9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.374096] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3785d233-c686-4bcc-868e-94acb61ca8a4 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "5ed9cee5-8cc8-4b70-96a9-737019363638" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.917s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.381238] env[62368]: DEBUG oslo_vmware.api [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 862.381238] env[62368]: value = "task-1198617" [ 862.381238] env[62368]: _type = "Task" [ 862.381238] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.391355] env[62368]: DEBUG oslo_vmware.api [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198617, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.395249] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8b00b0bc-775c-4d75-bf52-4a713b64d939 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.559s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.505907] env[62368]: DEBUG oslo_vmware.api [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198614, 'name': PowerOnVM_Task, 'duration_secs': 0.645924} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.506170] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 862.506445] env[62368]: INFO nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Took 9.13 seconds to spawn the instance on the hypervisor. [ 862.506719] env[62368]: DEBUG nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.507943] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab84859b-858b-4e94-bb79-39ae09e9554a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.524776] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.619608] env[62368]: DEBUG nova.scheduler.client.report [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.642166] env[62368]: DEBUG nova.network.neutron [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.701749] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198616, 'name': Rename_Task, 'duration_secs': 0.279554} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.702712] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 862.702712] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40c0264a-865d-4a84-a966-ace5001f5fae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.709527] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 862.709527] env[62368]: value = "task-1198618" [ 862.709527] env[62368]: _type = "Task" [ 862.709527] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.718263] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198618, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.730372] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.820111] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "5ed9cee5-8cc8-4b70-96a9-737019363638" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.820410] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "5ed9cee5-8cc8-4b70-96a9-737019363638" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.820623] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "5ed9cee5-8cc8-4b70-96a9-737019363638-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.820833] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "5ed9cee5-8cc8-4b70-96a9-737019363638-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.821076] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "5ed9cee5-8cc8-4b70-96a9-737019363638-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.823881] env[62368]: INFO nova.compute.manager [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Terminating instance [ 862.825885] env[62368]: DEBUG nova.compute.manager [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 862.826098] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.826979] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013cd4e9-fd5d-45bf-ac8e-8e2ace5f4c85 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.830537] env[62368]: DEBUG nova.network.neutron [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Updating instance_info_cache with network_info: [{"id": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "address": "fa:16:3e:fe:ae:1b", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b605458-f2", "ovs_interfaceid": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.838062] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.838363] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0cc10c40-5ef4-4d8a-a09f-fd6fad20359f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.846234] env[62368]: DEBUG oslo_vmware.api [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 862.846234] env[62368]: value = "task-1198619" [ 862.846234] env[62368]: _type = "Task" [ 862.846234] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.859516] env[62368]: DEBUG oslo_vmware.api [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.894448] env[62368]: DEBUG oslo_vmware.api [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198617, 'name': ReconfigVM_Task, 'duration_secs': 0.190486} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.894448] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259829', 'volume_id': '92efc3b4-cdd3-4084-8734-f57f630ab2db', 'name': 'volume-92efc3b4-cdd3-4084-8734-f57f630ab2db', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5554732b-34a1-41bd-9d0d-9bd39d62bec3', 'attached_at': '', 'detached_at': '', 'volume_id': '92efc3b4-cdd3-4084-8734-f57f630ab2db', 'serial': '92efc3b4-cdd3-4084-8734-f57f630ab2db'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 863.031220] env[62368]: INFO nova.compute.manager [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Took 38.30 seconds to build instance. [ 863.073244] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "58efc1cf-2469-41be-bf99-fc7b8c72113c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.073244] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "58efc1cf-2469-41be-bf99-fc7b8c72113c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.073244] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "58efc1cf-2469-41be-bf99-fc7b8c72113c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.073432] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "58efc1cf-2469-41be-bf99-fc7b8c72113c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.073513] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "58efc1cf-2469-41be-bf99-fc7b8c72113c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.077368] env[62368]: INFO nova.compute.manager [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Terminating instance [ 863.078574] env[62368]: DEBUG nova.compute.manager [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.078791] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 863.079611] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210d5b3f-a8be-400c-8177-ee509ce99819 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.094071] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 863.094408] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4969ca1-4ff2-4272-a9d4-cdf028c70662 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.103519] env[62368]: DEBUG nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 863.106845] env[62368]: DEBUG oslo_vmware.api [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 863.106845] env[62368]: value = "task-1198620" [ 863.106845] env[62368]: _type = "Task" [ 863.106845] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.117970] env[62368]: DEBUG oslo_vmware.api [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198620, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.127705] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.052s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.133805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.698s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.134266] env[62368]: DEBUG nova.objects.instance [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lazy-loading 'resources' on Instance uuid 29d2f0d6-7a76-4039-acbb-a3abd69d8370 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.149756] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.150057] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.150152] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.150341] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.150494] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.150651] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.150877] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.151051] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.151287] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.151397] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.151571] env[62368]: DEBUG nova.virt.hardware [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.153116] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd80f47-5b5c-49e5-9e0e-60332731d90c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.156916] env[62368]: INFO nova.scheduler.client.report [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleted allocations for instance d865d433-2341-4f6c-b840-609530ae2e51 [ 863.166548] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0548df-e119-4ef0-a6a2-fcbe0637e951 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.190951] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 863.191449] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259834', 'volume_id': '11c38512-c804-4553-9c51-d6a79caa8f45', 'name': 'volume-11c38512-c804-4553-9c51-d6a79caa8f45', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '58157ab1-80a4-427c-812b-f6fde1f8db68', 'attached_at': '', 'detached_at': '', 'volume_id': '11c38512-c804-4553-9c51-d6a79caa8f45', 'serial': '11c38512-c804-4553-9c51-d6a79caa8f45'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 863.192211] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c582ae-5c25-4c34-b01a-07f83b4d82af {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.214905] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284ae90e-2c12-43b1-ba7c-6f7f98f2b942 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.250526] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] volume-11c38512-c804-4553-9c51-d6a79caa8f45/volume-11c38512-c804-4553-9c51-d6a79caa8f45.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.250889] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198618, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.253418] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fd6a416-150f-48c9-8cd7-7486a10c7a00 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.273036] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 863.273036] env[62368]: value = "task-1198621" [ 863.273036] env[62368]: _type = "Task" [ 863.273036] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.283159] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198621, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.335149] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.335149] env[62368]: DEBUG nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Instance network_info: |[{"id": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "address": "fa:16:3e:fe:ae:1b", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b605458-f2", "ovs_interfaceid": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.335149] env[62368]: DEBUG oslo_concurrency.lockutils [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] Acquired lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.335149] env[62368]: DEBUG nova.network.neutron [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Refreshing network info cache for port 3b605458-f2bc-4c5e-8d9a-44fd80ab88bf {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 863.336535] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:ae:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b605458-f2bc-4c5e-8d9a-44fd80ab88bf', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.347197] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Creating folder: Project (3940819fc14a4c628aacd8820efe084e). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 863.348794] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2f04703-d5dd-4eb2-b0b7-6599d2fa061e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.360625] env[62368]: DEBUG oslo_vmware.api [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198619, 'name': PowerOffVM_Task, 'duration_secs': 0.222718} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.360875] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 863.361074] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 863.361705] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90661d7b-e834-4129-b01f-4e4b80a94781 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.365642] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Created folder: Project (3940819fc14a4c628aacd8820efe084e) in parent group-v259706. [ 863.365890] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Creating folder: Instances. Parent ref: group-v259835. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 863.366174] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-38e5229a-7151-4188-b310-0b0a02b9050c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.379200] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Created folder: Instances in parent group-v259835. [ 863.379661] env[62368]: DEBUG oslo.service.loopingcall [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.379800] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 863.380059] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0a19c7a-cede-4e90-9e50-cbafaa944f7a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.404936] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.404936] env[62368]: value = "task-1198625" [ 863.404936] env[62368]: _type = "Task" [ 863.404936] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.418046] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198625, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.446558] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.446952] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.447261] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleting the datastore file [datastore2] 5ed9cee5-8cc8-4b70-96a9-737019363638 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.447710] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d7353b3-4ad6-40b9-b728-998a272d201b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.459355] env[62368]: DEBUG oslo_vmware.api [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 863.459355] env[62368]: value = "task-1198626" [ 863.459355] env[62368]: _type = "Task" [ 863.459355] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.470287] env[62368]: DEBUG oslo_vmware.api [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198626, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.537374] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f50674f3-787d-4af4-b67b-6dd41824540b tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.994s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.569968] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.570791] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.570791] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.571011] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.571110] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.574556] env[62368]: INFO nova.compute.manager [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Terminating instance [ 863.576755] env[62368]: DEBUG nova.compute.manager [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.577052] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 863.578201] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315a65d3-5f8a-4495-a151-a98e97ed3af2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.589913] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 863.590335] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-676ac6aa-8c28-48c6-af89-2f314453cf0b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.601313] env[62368]: DEBUG oslo_vmware.api [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 863.601313] env[62368]: value = "task-1198627" [ 863.601313] env[62368]: _type = "Task" [ 863.601313] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.623532] env[62368]: DEBUG oslo_vmware.api [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198620, 'name': PowerOffVM_Task, 'duration_secs': 0.200035} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.627913] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 863.628274] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 863.628718] env[62368]: DEBUG oslo_vmware.api [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.630417] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21d52303-6f33-4f0a-b20e-b9fb50e395f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.637424] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "5af32858-dc9a-4380-827a-daf384c29bed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.637795] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.669046] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aa7f3806-549d-4431-87eb-3f11e48d58c7 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "d865d433-2341-4f6c-b840-609530ae2e51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.706s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.735966] env[62368]: DEBUG oslo_vmware.api [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198618, 'name': PowerOnVM_Task, 'duration_secs': 0.715417} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.736474] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 863.736829] env[62368]: INFO nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Took 7.19 seconds to spawn the instance on the hypervisor. [ 863.737061] env[62368]: DEBUG nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.738538] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b25c29-2a68-4f1f-91d6-86f4116a6d57 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.771024] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.771175] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.771456] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleting the datastore file [datastore2] 58efc1cf-2469-41be-bf99-fc7b8c72113c {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.771839] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e38e0483-45ec-406c-882c-5e6bbfd0f78b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.788188] env[62368]: DEBUG oslo_vmware.api [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 863.788188] env[62368]: value = "task-1198629" [ 863.788188] env[62368]: _type = "Task" [ 863.788188] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.798309] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198621, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.808922] env[62368]: DEBUG oslo_vmware.api [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198629, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.841966] env[62368]: DEBUG nova.network.neutron [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Successfully updated port: cb4b1e4c-15e5-4641-b434-e05e31283e5c {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.916440] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198625, 'name': CreateVM_Task, 'duration_secs': 0.449757} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.916594] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 863.919642] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.920050] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.920148] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.920746] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0679d60d-d9a2-45bb-bfea-7f941e8df6ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.926099] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 863.926099] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5266fb20-8be5-271e-e6c9-aca3314a3125" [ 863.926099] env[62368]: _type = "Task" [ 863.926099] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.939057] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5266fb20-8be5-271e-e6c9-aca3314a3125, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.944352] env[62368]: DEBUG nova.objects.instance [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'flavor' on Instance uuid 5554732b-34a1-41bd-9d0d-9bd39d62bec3 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.971797] env[62368]: DEBUG oslo_vmware.api [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198626, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182792} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.974699] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.974947] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 863.975159] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 863.979018] env[62368]: INFO nova.compute.manager [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Took 1.15 seconds to destroy the instance on the hypervisor. [ 863.979018] env[62368]: DEBUG oslo.service.loopingcall [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.979018] env[62368]: DEBUG nova.compute.manager [-] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.979018] env[62368]: DEBUG nova.network.neutron [-] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.031048] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d1da40-e080-4adf-8e71-76845257c88f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.039691] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd33f65-baed-447a-b0d1-3a22d0f85ae6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.078755] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3fb538-18fb-49a2-babd-75826efc5aaa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.088582] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a75ed3-0d82-4a07-8c30-a26a3f6390d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.111967] env[62368]: DEBUG nova.compute.provider_tree [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.124111] env[62368]: DEBUG oslo_vmware.api [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198627, 'name': PowerOffVM_Task, 'duration_secs': 0.252122} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.125240] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 864.125423] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 864.125977] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80aac017-3727-41c8-a951-62eacd7d3dd5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.146114] env[62368]: DEBUG nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.216653] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 864.217066] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 864.217385] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleting the datastore file [datastore2] f397e102-bfbd-4753-b1a4-4c5e9126b2c6 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.221235] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5cf9d707-dcd7-4b8c-b6bb-d3e9a261778b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.231892] env[62368]: DEBUG oslo_vmware.api [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 864.231892] env[62368]: value = "task-1198631" [ 864.231892] env[62368]: _type = "Task" [ 864.231892] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.245900] env[62368]: DEBUG oslo_vmware.api [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.274557] env[62368]: INFO nova.compute.manager [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Took 38.67 seconds to build instance. [ 864.287061] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198621, 'name': ReconfigVM_Task, 'duration_secs': 0.571548} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.287529] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Reconfigured VM instance instance-00000041 to attach disk [datastore1] volume-11c38512-c804-4553-9c51-d6a79caa8f45/volume-11c38512-c804-4553-9c51-d6a79caa8f45.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.294900] env[62368]: DEBUG nova.network.neutron [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Updated VIF entry in instance network info cache for port 3b605458-f2bc-4c5e-8d9a-44fd80ab88bf. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 864.295379] env[62368]: DEBUG nova.network.neutron [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Updating instance_info_cache with network_info: [{"id": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "address": "fa:16:3e:fe:ae:1b", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b605458-f2", "ovs_interfaceid": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.296709] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd9e9f81-7ae5-4991-8a12-58d81ff63686 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.308786] env[62368]: DEBUG nova.compute.manager [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Received event network-vif-plugged-cb4b1e4c-15e5-4641-b434-e05e31283e5c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.309053] env[62368]: DEBUG oslo_concurrency.lockutils [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] Acquiring lock "69ddb565-6c79-44e5-a7d1-d339ab426fae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.309292] env[62368]: DEBUG oslo_concurrency.lockutils [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] Lock "69ddb565-6c79-44e5-a7d1-d339ab426fae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.309467] env[62368]: DEBUG oslo_concurrency.lockutils [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] Lock "69ddb565-6c79-44e5-a7d1-d339ab426fae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.310039] env[62368]: DEBUG nova.compute.manager [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] No waiting events found dispatching network-vif-plugged-cb4b1e4c-15e5-4641-b434-e05e31283e5c {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 864.310039] env[62368]: WARNING nova.compute.manager [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Received unexpected event network-vif-plugged-cb4b1e4c-15e5-4641-b434-e05e31283e5c for instance with vm_state building and task_state spawning. [ 864.310039] env[62368]: DEBUG nova.compute.manager [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Received event network-changed-cb4b1e4c-15e5-4641-b434-e05e31283e5c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.310204] env[62368]: DEBUG nova.compute.manager [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Refreshing instance network info cache due to event network-changed-cb4b1e4c-15e5-4641-b434-e05e31283e5c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.310393] env[62368]: DEBUG oslo_concurrency.lockutils [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] Acquiring lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.310565] env[62368]: DEBUG oslo_concurrency.lockutils [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] Acquired lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.310783] env[62368]: DEBUG nova.network.neutron [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Refreshing network info cache for port cb4b1e4c-15e5-4641-b434-e05e31283e5c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.326766] env[62368]: DEBUG oslo_vmware.api [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198629, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.309815} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.328231] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.328490] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 864.328634] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 864.328831] env[62368]: INFO nova.compute.manager [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Took 1.25 seconds to destroy the instance on the hypervisor. [ 864.329191] env[62368]: DEBUG oslo.service.loopingcall [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.329772] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 864.329772] env[62368]: value = "task-1198632" [ 864.329772] env[62368]: _type = "Task" [ 864.329772] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.329962] env[62368]: DEBUG nova.compute.manager [-] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.330074] env[62368]: DEBUG nova.network.neutron [-] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.344789] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.345107] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198632, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.438306] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5266fb20-8be5-271e-e6c9-aca3314a3125, 'name': SearchDatastore_Task, 'duration_secs': 0.013433} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.438488] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.438724] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.438953] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.439120] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.439304] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.439579] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddf1a6bc-c6f2-4769-b1f5-b0b52786dd8f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.451073] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.451073] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 864.452351] env[62368]: DEBUG oslo_concurrency.lockutils [None req-adfaa3c1-2061-4dd7-bfbe-45d8db5be4f0 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.896s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.453984] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bf8048b-b443-464c-bb1c-d1e6954ade12 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.462197] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 864.462197] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52a326b3-1a1b-16a7-10a2-726b39b4bb12" [ 864.462197] env[62368]: _type = "Task" [ 864.462197] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.471801] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a326b3-1a1b-16a7-10a2-726b39b4bb12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.605299] env[62368]: DEBUG nova.compute.manager [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Stashing vm_state: stopped {{(pid=62368) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 864.619488] env[62368]: DEBUG nova.scheduler.client.report [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.666666] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.743672] env[62368]: DEBUG oslo_vmware.api [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198631, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325188} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.743982] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.744193] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 864.744376] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 864.744628] env[62368]: INFO nova.compute.manager [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Took 1.17 seconds to destroy the instance on the hypervisor. [ 864.744885] env[62368]: DEBUG oslo.service.loopingcall [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.745097] env[62368]: DEBUG nova.compute.manager [-] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.745193] env[62368]: DEBUG nova.network.neutron [-] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.780179] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e582e765-c821-45be-88e3-a2afd0bc0134 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "0cb0097b-2908-4107-bb57-0241915ccff7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.186s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.813581] env[62368]: DEBUG oslo_concurrency.lockutils [req-0522522a-68b1-4676-bbe1-31fcb938f869 req-b95f6402-7150-46c9-b246-358abef3323a service nova] Releasing lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.842588] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198632, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.872668] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "207de202-ca50-4811-84c5-a0d4454d29f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.872908] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "207de202-ca50-4811-84c5-a0d4454d29f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.875366] env[62368]: DEBUG nova.network.neutron [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 864.972931] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a326b3-1a1b-16a7-10a2-726b39b4bb12, 'name': SearchDatastore_Task, 'duration_secs': 0.012922} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.974241] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2f34b26-2762-487b-8810-cc71cd7f2523 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.980512] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 864.980512] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fe1433-8712-42a3-e265-7980c330245d" [ 864.980512] env[62368]: _type = "Task" [ 864.980512] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.989691] env[62368]: DEBUG nova.network.neutron [-] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.990897] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fe1433-8712-42a3-e265-7980c330245d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.038112] env[62368]: DEBUG nova.network.neutron [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.123926] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.127081] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.968s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.127081] env[62368]: DEBUG nova.objects.instance [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'resources' on Instance uuid 67d8c43a-d3af-407b-847a-1c2c8c53dbfb {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.128933] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.151025] env[62368]: INFO nova.scheduler.client.report [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleted allocations for instance 29d2f0d6-7a76-4039-acbb-a3abd69d8370 [ 865.190309] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.190598] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.190845] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.191055] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.191236] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.193747] env[62368]: INFO nova.compute.manager [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Terminating instance [ 865.197931] env[62368]: DEBUG nova.compute.manager [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 865.197931] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 865.197931] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e2dde6a-86a6-42d1-9eb1-a3e1677d78d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.205937] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 865.205937] env[62368]: value = "task-1198633" [ 865.205937] env[62368]: _type = "Task" [ 865.205937] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.212902] env[62368]: DEBUG nova.network.neutron [-] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.217826] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.343629] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198632, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.375347] env[62368]: DEBUG nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 865.464163] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "0cb0097b-2908-4107-bb57-0241915ccff7" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.464535] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "0cb0097b-2908-4107-bb57-0241915ccff7" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.464835] env[62368]: INFO nova.compute.manager [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Rebooting instance [ 865.492030] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fe1433-8712-42a3-e265-7980c330245d, 'name': SearchDatastore_Task, 'duration_secs': 0.036396} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.492218] env[62368]: INFO nova.compute.manager [-] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Took 1.52 seconds to deallocate network for instance. [ 865.492481] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.492736] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b87094da-6258-469e-ab37-5557955ad3a7/b87094da-6258-469e-ab37-5557955ad3a7.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 865.494536] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2eb3f744-0abb-49b1-9366-8788d7f89dbb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.504934] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 865.504934] env[62368]: value = "task-1198634" [ 865.504934] env[62368]: _type = "Task" [ 865.504934] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.513022] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198634, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.540202] env[62368]: DEBUG oslo_concurrency.lockutils [req-a066ed54-4d90-4490-ac0c-aedb23315ac2 req-e79aa423-c401-4811-bb79-53c9fe49989f service nova] Releasing lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.540705] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.540939] env[62368]: DEBUG nova.network.neutron [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 865.615117] env[62368]: DEBUG nova.network.neutron [-] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.658654] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d1ffdc23-2cf9-4947-bb1a-4e7633df57c8 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "29d2f0d6-7a76-4039-acbb-a3abd69d8370" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.217s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.720570] env[62368]: INFO nova.compute.manager [-] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Took 1.39 seconds to deallocate network for instance. [ 865.720661] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198633, 'name': PowerOffVM_Task, 'duration_secs': 0.422357} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.725575] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 865.726442] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 865.726541] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259829', 'volume_id': '92efc3b4-cdd3-4084-8734-f57f630ab2db', 'name': 'volume-92efc3b4-cdd3-4084-8734-f57f630ab2db', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5554732b-34a1-41bd-9d0d-9bd39d62bec3', 'attached_at': '', 'detached_at': '', 'volume_id': '92efc3b4-cdd3-4084-8734-f57f630ab2db', 'serial': '92efc3b4-cdd3-4084-8734-f57f630ab2db'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 865.729810] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9742c83-a908-4ac9-a67f-eefa33bf1227 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.756494] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac83f3e2-63a1-49bb-a4df-2ba61b3ab10e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.766021] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f46c90d-12ab-4a49-a837-07d9dd777ef4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.793199] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bab315-c2d7-401f-a463-b8bb0b1a778b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.810944] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] The volume has not been displaced from its original location: [datastore2] volume-92efc3b4-cdd3-4084-8734-f57f630ab2db/volume-92efc3b4-cdd3-4084-8734-f57f630ab2db.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 865.817023] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 865.820029] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec82e4aa-ce08-47e6-8e2c-27b7938320a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.840613] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 865.840613] env[62368]: value = "task-1198635" [ 865.840613] env[62368]: _type = "Task" [ 865.840613] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.842858] env[62368]: DEBUG oslo_vmware.api [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198632, 'name': ReconfigVM_Task, 'duration_secs': 1.179068} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.848677] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259834', 'volume_id': '11c38512-c804-4553-9c51-d6a79caa8f45', 'name': 'volume-11c38512-c804-4553-9c51-d6a79caa8f45', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '58157ab1-80a4-427c-812b-f6fde1f8db68', 'attached_at': '', 'detached_at': '', 'volume_id': '11c38512-c804-4553-9c51-d6a79caa8f45', 'serial': '11c38512-c804-4553-9c51-d6a79caa8f45'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 865.855707] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198635, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.893107] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.991852] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.991852] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquired lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.991999] env[62368]: DEBUG nova.network.neutron [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 866.001345] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.019812] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198634, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.038803] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a50c57e-c3d4-43b1-b9ee-18c7bf2d9e69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.054201] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8657e57-6214-4528-9527-524830ed534e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.091164] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207c455d-4a93-42a3-a6e3-f96844b82a64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.093886] env[62368]: DEBUG nova.network.neutron [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 866.104184] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b83b315-0141-4189-ab3a-65945e526ab1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.123796] env[62368]: INFO nova.compute.manager [-] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Took 1.38 seconds to deallocate network for instance. [ 866.124472] env[62368]: DEBUG nova.compute.provider_tree [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.226733] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.268652] env[62368]: DEBUG nova.network.neutron [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Updating instance_info_cache with network_info: [{"id": "cb4b1e4c-15e5-4641-b434-e05e31283e5c", "address": "fa:16:3e:22:8c:36", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb4b1e4c-15", "ovs_interfaceid": "cb4b1e4c-15e5-4641-b434-e05e31283e5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.356630] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198635, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.405420] env[62368]: DEBUG nova.compute.manager [req-2849acbf-2949-4818-a815-91c3b56b9229 req-5894bb25-d580-4f70-ab24-686d8e6fedca service nova] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Received event network-vif-deleted-b6d50c17-ef64-4171-8459-4829ffa5d03d {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.405813] env[62368]: DEBUG nova.compute.manager [req-2849acbf-2949-4818-a815-91c3b56b9229 req-5894bb25-d580-4f70-ab24-686d8e6fedca service nova] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Received event network-vif-deleted-bd3e0156-36ff-4c90-8484-69fed4dfa149 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.406125] env[62368]: DEBUG nova.compute.manager [req-2849acbf-2949-4818-a815-91c3b56b9229 req-5894bb25-d580-4f70-ab24-686d8e6fedca service nova] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Received event network-vif-deleted-d6d93849-79aa-4366-a8b1-ce1732983fa4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.516579] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198634, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.759848} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.516829] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b87094da-6258-469e-ab37-5557955ad3a7/b87094da-6258-469e-ab37-5557955ad3a7.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 866.517151] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.517348] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84d1d9fb-1d13-4341-a432-aca945a97932 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.528121] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 866.528121] env[62368]: value = "task-1198636" [ 866.528121] env[62368]: _type = "Task" [ 866.528121] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.537172] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.627334] env[62368]: DEBUG nova.scheduler.client.report [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.637386] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.770539] env[62368]: DEBUG nova.network.neutron [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Updating instance_info_cache with network_info: [{"id": "b0fe605d-e7f1-4955-9532-b1e6aa918864", "address": "fa:16:3e:9e:4a:a2", "network": {"id": "2038d073-bf70-4f85-adfa-cf1d464d908a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-561693200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da6a595e1dc64441a999ace8ebb22151", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc6e6fe1-c4f5-4389-a49f-0978060eebb4", "external-id": "nsx-vlan-transportzone-829", "segmentation_id": 829, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0fe605d-e7", "ovs_interfaceid": "b0fe605d-e7f1-4955-9532-b1e6aa918864", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.772948] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.773329] env[62368]: DEBUG nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Instance network_info: |[{"id": "cb4b1e4c-15e5-4641-b434-e05e31283e5c", "address": "fa:16:3e:22:8c:36", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb4b1e4c-15", "ovs_interfaceid": "cb4b1e4c-15e5-4641-b434-e05e31283e5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 866.774124] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:8c:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb4b1e4c-15e5-4641-b434-e05e31283e5c', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.788155] env[62368]: DEBUG oslo.service.loopingcall [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.788775] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 866.788952] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a00563e-6e70-4486-a09d-4be06d92315e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.820474] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.820474] env[62368]: value = "task-1198637" [ 866.820474] env[62368]: _type = "Task" [ 866.820474] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.829798] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198637, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.853515] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198635, 'name': ReconfigVM_Task, 'duration_secs': 0.803581} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.853651] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 866.858489] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fe9b7df-1779-42cb-bad4-8eca46fe5a91 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.877049] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 866.877049] env[62368]: value = "task-1198638" [ 866.877049] env[62368]: _type = "Task" [ 866.877049] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.891861] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198638, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.895716] env[62368]: DEBUG nova.objects.instance [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.041957] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077693} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.041957] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.043368] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6634da01-b196-4336-a544-b7ed3b3683a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.072352] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] b87094da-6258-469e-ab37-5557955ad3a7/b87094da-6258-469e-ab37-5557955ad3a7.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.072718] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77d8ef26-0980-4212-a857-ab547491e21d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.095327] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 867.095327] env[62368]: value = "task-1198639" [ 867.095327] env[62368]: _type = "Task" [ 867.095327] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.107513] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198639, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.121386] env[62368]: DEBUG oslo_concurrency.lockutils [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.133572] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.007s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.136791] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 10.584s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.163641] env[62368]: INFO nova.scheduler.client.report [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleted allocations for instance 67d8c43a-d3af-407b-847a-1c2c8c53dbfb [ 867.275299] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Releasing lock "refresh_cache-0cb0097b-2908-4107-bb57-0241915ccff7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.277850] env[62368]: DEBUG nova.compute.manager [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.279058] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b8baf6-1bcb-4a64-a23e-9859b1f4cc32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.331690] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198637, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.376217] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.376580] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.377653] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.377653] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.377653] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.379616] env[62368]: INFO nova.compute.manager [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Terminating instance [ 867.381599] env[62368]: DEBUG nova.compute.manager [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 867.381812] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 867.382737] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bd90da-8397-4487-8809-5e76370a52a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.395570] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198638, 'name': ReconfigVM_Task, 'duration_secs': 0.308276} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.397899] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259829', 'volume_id': '92efc3b4-cdd3-4084-8734-f57f630ab2db', 'name': 'volume-92efc3b4-cdd3-4084-8734-f57f630ab2db', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5554732b-34a1-41bd-9d0d-9bd39d62bec3', 'attached_at': '', 'detached_at': '', 'volume_id': '92efc3b4-cdd3-4084-8734-f57f630ab2db', 'serial': '92efc3b4-cdd3-4084-8734-f57f630ab2db'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 867.398720] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 867.398720] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 867.401221] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298b364c-3023-4640-a5ef-cc081a7519dd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.404403] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d2df1a4-1e8e-413e-83cf-cd12108bfea7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.406320] env[62368]: DEBUG oslo_concurrency.lockutils [None req-525e11a8-c985-4a03-8890-5f4b925c9f47 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.842s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.408091] env[62368]: DEBUG oslo_concurrency.lockutils [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.287s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.408499] env[62368]: DEBUG nova.compute.manager [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.410900] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb16e77f-8e22-4247-8e4a-6531bf05a5f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.418518] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 867.419031] env[62368]: DEBUG oslo_vmware.api [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 867.419031] env[62368]: value = "task-1198640" [ 867.419031] env[62368]: _type = "Task" [ 867.419031] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.419595] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8df75e77-39b8-446c-ab3f-4493a1d4bb48 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.430815] env[62368]: DEBUG nova.compute.manager [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62368) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 867.431822] env[62368]: DEBUG nova.objects.instance [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.439852] env[62368]: DEBUG oslo_vmware.api [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198640, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.453916] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.454219] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.454486] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.454820] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.454985] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.457498] env[62368]: INFO nova.compute.manager [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Terminating instance [ 867.459566] env[62368]: DEBUG nova.compute.manager [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 867.459827] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 867.460670] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3097246-eb4a-40e5-9a46-39a363254118 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.469640] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 867.469998] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f17826c4-406f-4d29-bbda-8603c6cb7002 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.477532] env[62368]: DEBUG oslo_vmware.api [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 867.477532] env[62368]: value = "task-1198642" [ 867.477532] env[62368]: _type = "Task" [ 867.477532] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.487580] env[62368]: DEBUG oslo_vmware.api [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198642, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.517684] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 867.517984] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 867.518232] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleting the datastore file [datastore1] 5554732b-34a1-41bd-9d0d-9bd39d62bec3 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.518567] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75003e31-6dca-47c4-aafc-031e4aff146f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.526141] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 867.526141] env[62368]: value = "task-1198643" [ 867.526141] env[62368]: _type = "Task" [ 867.526141] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.535015] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198643, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.606669] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198639, 'name': ReconfigVM_Task, 'duration_secs': 0.372577} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.607162] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Reconfigured VM instance instance-0000004e to attach disk [datastore1] b87094da-6258-469e-ab37-5557955ad3a7/b87094da-6258-469e-ab37-5557955ad3a7.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.607873] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e47582ee-1526-4b95-984c-5429175a4d2e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.618808] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 867.618808] env[62368]: value = "task-1198644" [ 867.618808] env[62368]: _type = "Task" [ 867.618808] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.628935] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198644, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.640825] env[62368]: DEBUG nova.objects.instance [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lazy-loading 'migration_context' on Instance uuid a0cea538-b162-4504-ac34-803a2d5a8071 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.671448] env[62368]: DEBUG oslo_concurrency.lockutils [None req-426d380d-244e-4502-90a6-e422d1c30bf9 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.712s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.672521] env[62368]: DEBUG oslo_concurrency.lockutils [req-5dcb85ab-58b6-464e-ab96-00f27212f8a8 req-4757df3f-68c3-473e-a4d1-200144a83bcf service nova] Acquired lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.673436] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5e3a8b-a857-4c90-9cf4-caa8161e3b08 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.683469] env[62368]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 867.683648] env[62368]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62368) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 867.684416] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-150b67dc-5a0e-4502-8375-0374b751f705 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.695230] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4687818d-c8d2-470c-9d37-323b305eab6c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.728023] env[62368]: ERROR root [req-5dcb85ab-58b6-464e-ab96-00f27212f8a8 req-4757df3f-68c3-473e-a4d1-200144a83bcf service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-259777' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-259777' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-259777' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-259777'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-259777' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-259777' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-259777'}\n"]: nova.exception.InstanceNotFound: Instance 67d8c43a-d3af-407b-847a-1c2c8c53dbfb could not be found. [ 867.728204] env[62368]: DEBUG oslo_concurrency.lockutils [req-5dcb85ab-58b6-464e-ab96-00f27212f8a8 req-4757df3f-68c3-473e-a4d1-200144a83bcf service nova] Releasing lock "67d8c43a-d3af-407b-847a-1c2c8c53dbfb" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.728433] env[62368]: DEBUG nova.compute.manager [req-5dcb85ab-58b6-464e-ab96-00f27212f8a8 req-4757df3f-68c3-473e-a4d1-200144a83bcf service nova] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Detach interface failed, port_id=c582c777-e544-40af-9f04-d8c171475bc1, reason: Instance 67d8c43a-d3af-407b-847a-1c2c8c53dbfb could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 867.832020] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198637, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.933244] env[62368]: DEBUG oslo_vmware.api [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198640, 'name': PowerOffVM_Task, 'duration_secs': 0.231601} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.935426] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 867.935625] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 867.936013] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14cf1993-5b9e-479f-8e96-5b79fe6efbff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.938975] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 867.939221] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1f81603-c8af-4d0f-b692-9ecde516da35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.947397] env[62368]: DEBUG oslo_vmware.api [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 867.947397] env[62368]: value = "task-1198645" [ 867.947397] env[62368]: _type = "Task" [ 867.947397] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.958561] env[62368]: DEBUG oslo_vmware.api [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198645, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.987689] env[62368]: DEBUG oslo_vmware.api [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198642, 'name': PowerOffVM_Task, 'duration_secs': 0.168634} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.988123] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 867.988228] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 867.988452] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96b9914d-e0e8-46b7-8a60-e9c33cd4d84a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.017171] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 868.017516] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 868.017738] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleting the datastore file [datastore2] cda0cf42-11ae-4da3-b838-5bac4bc2f4d1 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.018069] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de92e11d-5950-4451-a876-0dc6def3f55c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.025902] env[62368]: DEBUG oslo_vmware.api [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 868.025902] env[62368]: value = "task-1198648" [ 868.025902] env[62368]: _type = "Task" [ 868.025902] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.037383] env[62368]: DEBUG oslo_vmware.api [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198648, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.040608] env[62368]: DEBUG oslo_vmware.api [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198643, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163212} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.040878] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.041040] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 868.041226] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 868.041405] env[62368]: INFO nova.compute.manager [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Took 2.85 seconds to destroy the instance on the hypervisor. [ 868.041650] env[62368]: DEBUG oslo.service.loopingcall [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.041849] env[62368]: DEBUG nova.compute.manager [-] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 868.041945] env[62368]: DEBUG nova.network.neutron [-] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.058539] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 868.058768] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 868.058957] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleting the datastore file [datastore2] e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.059273] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2dd7deb3-63cf-4719-a3f8-e53b207fedfb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.067447] env[62368]: DEBUG oslo_vmware.api [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for the task: (returnval){ [ 868.067447] env[62368]: value = "task-1198649" [ 868.067447] env[62368]: _type = "Task" [ 868.067447] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.077984] env[62368]: DEBUG oslo_vmware.api [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.132474] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198644, 'name': Rename_Task, 'duration_secs': 0.147969} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.132890] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 868.133264] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d423d34-f433-4f0a-8027-5f323dd7a80b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.147356] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 868.147356] env[62368]: value = "task-1198650" [ 868.147356] env[62368]: _type = "Task" [ 868.147356] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.169848] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198650, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.296949] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d96526e-6231-435a-a3f7-c4fd8e961c3c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.309861] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Doing hard reboot of VM {{(pid=62368) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 868.311029] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-d3b30eac-02f3-4537-8ac8-b2daef116f39 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.318530] env[62368]: DEBUG oslo_vmware.api [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 868.318530] env[62368]: value = "task-1198651" [ 868.318530] env[62368]: _type = "Task" [ 868.318530] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.346146] env[62368]: DEBUG oslo_vmware.api [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198651, 'name': ResetVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.346644] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198637, 'name': CreateVM_Task, 'duration_secs': 1.422514} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.349177] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 868.350083] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.350292] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.350838] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.350924] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb586ca9-fd3f-43db-9f20-ec3999cb3bde {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.357187] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 868.357187] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52692c73-4691-b1c8-57f4-1a3249c89147" [ 868.357187] env[62368]: _type = "Task" [ 868.357187] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.371067] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52692c73-4691-b1c8-57f4-1a3249c89147, 'name': SearchDatastore_Task, 'duration_secs': 0.010892} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.371166] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.371369] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.371695] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.371773] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.372025] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.372321] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90b0afe5-888a-4b58-97ff-78a6a7d45adc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.382248] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.382468] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 868.383899] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85497647-96ce-4d01-b52e-ce3974cb3d01 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.389252] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 868.389252] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e29faf-8e0c-cad4-dd85-ac9474cc00f2" [ 868.389252] env[62368]: _type = "Task" [ 868.389252] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.401808] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e29faf-8e0c-cad4-dd85-ac9474cc00f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.458611] env[62368]: DEBUG oslo_vmware.api [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198645, 'name': PowerOffVM_Task, 'duration_secs': 0.232001} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.462019] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 868.462832] env[62368]: DEBUG nova.compute.manager [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.464185] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b371bb-746d-449c-b92a-7a9336390ab6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.532635] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f5a2af-2164-4034-8068-a0a5c7210010 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.547057] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eca9ed8-ca11-408b-9163-e0704acd5e73 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.550989] env[62368]: DEBUG oslo_vmware.api [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198648, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166912} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.551431] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.551900] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 868.552228] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 868.552615] env[62368]: INFO nova.compute.manager [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Took 1.17 seconds to destroy the instance on the hypervisor. [ 868.553034] env[62368]: DEBUG oslo.service.loopingcall [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.553748] env[62368]: DEBUG nova.compute.manager [-] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 868.553977] env[62368]: DEBUG nova.network.neutron [-] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.593285] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6e5703-e63f-4b0c-99a5-d3eea85ad8b2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.598124] env[62368]: DEBUG nova.compute.manager [req-1b1b4560-496d-45e3-8777-e53d64c2bfac req-28160275-6182-424d-a7f3-7efd37718e63 service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Received event network-vif-deleted-f2865835-685b-4772-8401-30b2455e11c8 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.598305] env[62368]: INFO nova.compute.manager [req-1b1b4560-496d-45e3-8777-e53d64c2bfac req-28160275-6182-424d-a7f3-7efd37718e63 service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Neutron deleted interface f2865835-685b-4772-8401-30b2455e11c8; detaching it from the instance and deleting it from the info cache [ 868.598491] env[62368]: DEBUG nova.network.neutron [req-1b1b4560-496d-45e3-8777-e53d64c2bfac req-28160275-6182-424d-a7f3-7efd37718e63 service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.611198] env[62368]: DEBUG oslo_vmware.api [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Task: {'id': task-1198649, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168845} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.614274] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.614590] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 868.614761] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 868.614948] env[62368]: INFO nova.compute.manager [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Took 1.16 seconds to destroy the instance on the hypervisor. [ 868.615579] env[62368]: DEBUG oslo.service.loopingcall [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.616572] env[62368]: DEBUG nova.compute.manager [-] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 868.616731] env[62368]: DEBUG nova.network.neutron [-] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.619355] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b56e30e-2c34-4811-a0f3-51e4ebf7139a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.636789] env[62368]: DEBUG nova.compute.provider_tree [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.659848] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198650, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.833345] env[62368]: DEBUG oslo_vmware.api [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198651, 'name': ResetVM_Task, 'duration_secs': 0.095873} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.833756] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Did hard reboot of VM {{(pid=62368) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 868.833989] env[62368]: DEBUG nova.compute.manager [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.834864] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232e9631-21ba-497e-8191-7f72c06cc355 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.906616] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e29faf-8e0c-cad4-dd85-ac9474cc00f2, 'name': SearchDatastore_Task, 'duration_secs': 0.015999} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.907659] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-660666bd-d13c-4d1d-bba0-5dcbab6ff83a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.914278] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 868.914278] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52cb11a7-b191-543b-e73e-e965cfccbfb8" [ 868.914278] env[62368]: _type = "Task" [ 868.914278] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.926587] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cb11a7-b191-543b-e73e-e965cfccbfb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.979669] env[62368]: DEBUG oslo_concurrency.lockutils [None req-945aaf05-d78a-4ef6-a917-f642db9e050e tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.572s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.981222] env[62368]: DEBUG nova.network.neutron [-] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.105410] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f1a13d3-ca82-4cfb-9eaf-6379e56dacdd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.121895] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea79d5a-29da-4e3e-b944-e76de8bef7ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.140605] env[62368]: DEBUG nova.scheduler.client.report [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.165928] env[62368]: DEBUG nova.compute.manager [req-1b1b4560-496d-45e3-8777-e53d64c2bfac req-28160275-6182-424d-a7f3-7efd37718e63 service nova] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Detach interface failed, port_id=f2865835-685b-4772-8401-30b2455e11c8, reason: Instance 5554732b-34a1-41bd-9d0d-9bd39d62bec3 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 869.173155] env[62368]: DEBUG oslo_vmware.api [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198650, 'name': PowerOnVM_Task, 'duration_secs': 0.56865} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.173575] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 869.174861] env[62368]: INFO nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Took 8.63 seconds to spawn the instance on the hypervisor. [ 869.174861] env[62368]: DEBUG nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.175543] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d6e41a-9ace-4834-b587-66da0b0efbf9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.348773] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6ccfd429-72e6-4f40-9827-3e15d9cf6617 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "0cb0097b-2908-4107-bb57-0241915ccff7" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.884s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.425914] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cb11a7-b191-543b-e73e-e965cfccbfb8, 'name': SearchDatastore_Task, 'duration_secs': 0.025123} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.425914] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.426161] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae/69ddb565-6c79-44e5-a7d1-d339ab426fae.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 869.426431] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49939ecf-b684-458e-84af-cf3afb7969da {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.434438] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 869.434438] env[62368]: value = "task-1198652" [ 869.434438] env[62368]: _type = "Task" [ 869.434438] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.448375] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.473975] env[62368]: DEBUG nova.network.neutron [-] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.487564] env[62368]: INFO nova.compute.manager [-] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Took 1.44 seconds to deallocate network for instance. [ 869.542803] env[62368]: DEBUG nova.network.neutron [-] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.620956] env[62368]: DEBUG nova.objects.instance [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.695761] env[62368]: INFO nova.compute.manager [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Took 37.97 seconds to build instance. [ 869.946948] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198652, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.976907] env[62368]: INFO nova.compute.manager [-] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Took 1.42 seconds to deallocate network for instance. [ 870.035013] env[62368]: INFO nova.compute.manager [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Took 0.55 seconds to detach 1 volumes for instance. [ 870.046896] env[62368]: INFO nova.compute.manager [-] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Took 1.43 seconds to deallocate network for instance. [ 870.127661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.127797] env[62368]: DEBUG oslo_concurrency.lockutils [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.128589] env[62368]: DEBUG nova.network.neutron [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 870.128752] env[62368]: DEBUG nova.objects.instance [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'info_cache' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.153177] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.016s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.163973] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.605s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.197947] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d5fd35c-a7ef-4284-9b47-871f47db7a77 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "b87094da-6258-469e-ab37-5557955ad3a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.489s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.303082] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "a5cbadbd-20dd-4514-8867-20243af5db0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.303381] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.447606] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198652, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527725} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.447873] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae/69ddb565-6c79-44e5-a7d1-d339ab426fae.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 870.448135] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.448364] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57ca9829-6c5b-428d-88f0-b84c07585e5f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.456065] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 870.456065] env[62368]: value = "task-1198653" [ 870.456065] env[62368]: _type = "Task" [ 870.456065] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.464992] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.485237] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.505941] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "0cb0097b-2908-4107-bb57-0241915ccff7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.506247] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "0cb0097b-2908-4107-bb57-0241915ccff7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.506483] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "0cb0097b-2908-4107-bb57-0241915ccff7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.506689] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "0cb0097b-2908-4107-bb57-0241915ccff7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.506949] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "0cb0097b-2908-4107-bb57-0241915ccff7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.509824] env[62368]: INFO nova.compute.manager [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Terminating instance [ 870.511722] env[62368]: DEBUG nova.compute.manager [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 870.511941] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 870.513193] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113e0cfc-b299-46a4-8bb4-86c0916364b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.522261] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 870.522525] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f254428-4d7e-4e57-9c55-2f4a5e610838 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.529295] env[62368]: DEBUG oslo_vmware.api [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 870.529295] env[62368]: value = "task-1198654" [ 870.529295] env[62368]: _type = "Task" [ 870.529295] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.538047] env[62368]: DEBUG oslo_vmware.api [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.541056] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.553046] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.619158] env[62368]: DEBUG nova.compute.manager [req-c30b8e33-e14e-441f-8bed-264bd037cf02 req-7ef0783e-c754-4844-80a8-17fb8a16b719 service nova] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Received event network-vif-deleted-07be57e1-db65-4ba2-a054-72ee59275498 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.619158] env[62368]: DEBUG nova.compute.manager [req-c30b8e33-e14e-441f-8bed-264bd037cf02 req-7ef0783e-c754-4844-80a8-17fb8a16b719 service nova] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Received event network-vif-deleted-b65f0e56-58ab-48a2-90ca-ca285f331170 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.633721] env[62368]: DEBUG nova.objects.base [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Object Instance<58157ab1-80a4-427c-812b-f6fde1f8db68> lazy-loaded attributes: flavor,info_cache {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 870.805703] env[62368]: DEBUG nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 870.966476] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067937} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.966946] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.967627] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298cb34f-3e45-4e0e-b665-d72783fb0cb1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.992976] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae/69ddb565-6c79-44e5-a7d1-d339ab426fae.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.993514] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f67d8fe-5382-439d-bdf2-82033958a758 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.014306] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 871.014306] env[62368]: value = "task-1198655" [ 871.014306] env[62368]: _type = "Task" [ 871.014306] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.023798] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198655, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.038912] env[62368]: DEBUG oslo_vmware.api [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.180172] env[62368]: INFO nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating resource usage from migration a4a55aca-f211-40de-8c6e-9835ad122ae2 [ 871.203573] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 417f000f-cf23-404d-877c-45990d1a7c77 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 871.203740] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b79f0e79-9e3a-47c7-9949-8743601ec6c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 871.203958] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance eea21546-fbbf-4440-829c-8583c4ccabb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 871.204047] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 58157ab1-80a4-427c-812b-f6fde1f8db68 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 871.204145] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance cda0cf42-11ae-4da3-b838-5bac4bc2f4d1 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 871.204273] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 871.204406] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 2a46c954-449e-4d62-be80-add1040ed4c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 871.204532] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 58efc1cf-2469-41be-bf99-fc7b8c72113c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 871.204706] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 5554732b-34a1-41bd-9d0d-9bd39d62bec3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 871.204843] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 871.330853] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.420492] env[62368]: DEBUG nova.network.neutron [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updating instance_info_cache with network_info: [{"id": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "address": "fa:16:3e:23:0d:cc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46fba643-18", "ovs_interfaceid": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.527799] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198655, 'name': ReconfigVM_Task, 'duration_secs': 0.303976} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.528114] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae/69ddb565-6c79-44e5-a7d1-d339ab426fae.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.528751] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4981f0bd-5cb2-4b29-a924-5a8dd5dad394 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.542031] env[62368]: DEBUG oslo_vmware.api [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198654, 'name': PowerOffVM_Task, 'duration_secs': 0.697139} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.543621] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 871.543923] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 871.544382] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 871.544382] env[62368]: value = "task-1198656" [ 871.544382] env[62368]: _type = "Task" [ 871.544382] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.544655] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b1c4c3a-a39b-4ede-a988-b5dc2007ad3c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.556212] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198656, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.662690] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 871.662690] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 871.662690] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Deleting the datastore file [datastore2] 0cb0097b-2908-4107-bb57-0241915ccff7 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.662977] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b07b718-db37-411c-80fa-6b46e027ba8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.671344] env[62368]: DEBUG oslo_vmware.api [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for the task: (returnval){ [ 871.671344] env[62368]: value = "task-1198658" [ 871.671344] env[62368]: _type = "Task" [ 871.671344] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.680559] env[62368]: DEBUG oslo_vmware.api [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.702909] env[62368]: INFO nova.compute.manager [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Swapping old allocation on dict_keys(['2202a74c-753d-4e1d-a031-7cefe24ee9d6']) held by migration 97c6a2e1-faef-43ad-8d88-091a4fcca87e for instance [ 871.707697] env[62368]: INFO nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 97c6a2e1-faef-43ad-8d88-091a4fcca87e has allocations against this compute host but is not found in the database. [ 871.707697] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a0cea538-b162-4504-ac34-803a2d5a8071 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 871.707839] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 5ed9cee5-8cc8-4b70-96a9-737019363638 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 871.708017] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance f397e102-bfbd-4753-b1a4-4c5e9126b2c6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 871.708174] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 0cb0097b-2908-4107-bb57-0241915ccff7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 871.708321] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b87094da-6258-469e-ab37-5557955ad3a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 871.708468] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 69ddb565-6c79-44e5-a7d1-d339ab426fae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 871.728460] env[62368]: DEBUG nova.scheduler.client.report [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Overwriting current allocation {'allocations': {'2202a74c-753d-4e1d-a031-7cefe24ee9d6': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 99}}, 'project_id': '514286cc83654bf2a7a01d6c5df2b195', 'user_id': '86d43d611b904b1c88e925db1e7f26ea', 'consumer_generation': 1} on consumer a0cea538-b162-4504-ac34-803a2d5a8071 {{(pid=62368) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 871.820517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.820709] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquired lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.820892] env[62368]: DEBUG nova.network.neutron [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 871.923418] env[62368]: DEBUG oslo_concurrency.lockutils [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Releasing lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.057482] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198656, 'name': Rename_Task, 'duration_secs': 0.151732} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.057843] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.058026] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91ac21cb-ef8e-491e-9f2e-517d330bcf38 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.065552] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 872.065552] env[62368]: value = "task-1198659" [ 872.065552] env[62368]: _type = "Task" [ 872.065552] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.073810] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198659, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.184224] env[62368]: DEBUG oslo_vmware.api [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Task: {'id': task-1198658, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139073} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.184538] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.184842] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 872.185120] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 872.185352] env[62368]: INFO nova.compute.manager [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Took 1.67 seconds to destroy the instance on the hypervisor. [ 872.185691] env[62368]: DEBUG oslo.service.loopingcall [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.185955] env[62368]: DEBUG nova.compute.manager [-] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 872.186090] env[62368]: DEBUG nova.network.neutron [-] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 872.211705] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 55eecf3d-501c-490d-bddd-0211fd082841 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 872.426800] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.427200] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea36144d-efb7-4a1a-a654-9ef54c713fba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.436626] env[62368]: DEBUG oslo_vmware.api [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 872.436626] env[62368]: value = "task-1198660" [ 872.436626] env[62368]: _type = "Task" [ 872.436626] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.445553] env[62368]: DEBUG oslo_vmware.api [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198660, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.578205] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198659, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.717541] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 5af32858-dc9a-4380-827a-daf384c29bed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 872.717710] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Migration a4a55aca-f211-40de-8c6e-9835ad122ae2 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 872.717836] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 13765305-2e55-4ee8-9a6f-4ae5ee724367 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 872.764711] env[62368]: DEBUG nova.compute.manager [req-869aa7bf-f4e5-4ca5-8667-1278400926af req-b4b2a8f4-ad50-477b-8c77-be4ce6d4618a service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Received event network-vif-deleted-b0fe605d-e7f1-4955-9532-b1e6aa918864 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.764957] env[62368]: INFO nova.compute.manager [req-869aa7bf-f4e5-4ca5-8667-1278400926af req-b4b2a8f4-ad50-477b-8c77-be4ce6d4618a service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Neutron deleted interface b0fe605d-e7f1-4955-9532-b1e6aa918864; detaching it from the instance and deleting it from the info cache [ 872.765165] env[62368]: DEBUG nova.network.neutron [req-869aa7bf-f4e5-4ca5-8667-1278400926af req-b4b2a8f4-ad50-477b-8c77-be4ce6d4618a service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.831087] env[62368]: DEBUG nova.network.neutron [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance_info_cache with network_info: [{"id": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "address": "fa:16:3e:98:a0:d2", "network": {"id": "34f156be-cab3-45a2-8b52-74dd4cd1e42d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2a054e975862463bb842479ad19393ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap444a47e3-2f", "ovs_interfaceid": "444a47e3-2f56-4895-8a8b-7e5605dbfaf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.946949] env[62368]: DEBUG oslo_vmware.api [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198660, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.077738] env[62368]: DEBUG oslo_vmware.api [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198659, 'name': PowerOnVM_Task, 'duration_secs': 0.521145} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.078050] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.078266] env[62368]: INFO nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Took 9.97 seconds to spawn the instance on the hypervisor. [ 873.078459] env[62368]: DEBUG nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.079268] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca18130-a58d-4158-8303-1cd20c90752f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.223077] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 207de202-ca50-4811-84c5-a0d4454d29f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 873.238053] env[62368]: DEBUG nova.network.neutron [-] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.267741] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b81b58c-8b3e-4d13-8104-d30b3e447a55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.277422] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952b6d9c-28b0-437a-ab19-5f02a1d043c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.307651] env[62368]: DEBUG nova.compute.manager [req-869aa7bf-f4e5-4ca5-8667-1278400926af req-b4b2a8f4-ad50-477b-8c77-be4ce6d4618a service nova] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Detach interface failed, port_id=b0fe605d-e7f1-4955-9532-b1e6aa918864, reason: Instance 0cb0097b-2908-4107-bb57-0241915ccff7 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 873.334244] env[62368]: DEBUG oslo_concurrency.lockutils [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Releasing lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.334804] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 873.335139] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9819d2a1-e228-4d91-9935-c40d6df403e3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.344037] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 873.344037] env[62368]: value = "task-1198661" [ 873.344037] env[62368]: _type = "Task" [ 873.344037] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.353701] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198661, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.449261] env[62368]: DEBUG oslo_vmware.api [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198660, 'name': PowerOnVM_Task, 'duration_secs': 0.822432} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.449563] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.449797] env[62368]: DEBUG nova.compute.manager [None req-664af16d-0cce-4f44-a6d6-567bca3a6b0b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.450680] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa2227c-4ded-4e17-b5c5-47b6be0da84f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.597280] env[62368]: INFO nova.compute.manager [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Took 39.48 seconds to build instance. [ 873.726180] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a5cbadbd-20dd-4514-8867-20243af5db0c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 873.726457] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 873.726607] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 873.740792] env[62368]: INFO nova.compute.manager [-] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Took 1.55 seconds to deallocate network for instance. [ 873.858336] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198661, 'name': PowerOffVM_Task, 'duration_secs': 0.217143} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.861052] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 873.861805] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:16:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f181c86a-16f3-4905-b3d9-783b0740eec7',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1115957553',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.862074] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.862294] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.862537] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.862765] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.862961] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.863252] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.863460] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.863675] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.863900] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.864142] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.870240] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa29d769-e67e-4040-bc17-a66503244969 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.888919] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 873.888919] env[62368]: value = "task-1198662" [ 873.888919] env[62368]: _type = "Task" [ 873.888919] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.900374] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198662, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.964352] env[62368]: INFO nova.compute.manager [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Rescuing [ 873.964352] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.964352] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.964352] env[62368]: DEBUG nova.network.neutron [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.018348] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46447de-efa4-49e6-8d29-3c24dbe17e53 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.028102] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43babaf2-93d3-4b2c-a31a-ce6de7222976 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.058582] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421294ca-9f20-4718-be02-510c518b3368 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.066619] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cc771a-8435-4928-b305-161aaf28382a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.080174] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.099348] env[62368]: DEBUG oslo_concurrency.lockutils [None req-31b01641-cd27-4407-87b7-b93900993722 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "69ddb565-6c79-44e5-a7d1-d339ab426fae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.706s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.247485] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.400504] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198662, 'name': ReconfigVM_Task, 'duration_secs': 0.140259} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.402058] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461d9293-2814-43d8-bd0d-2c566853769f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.419686] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:16:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f181c86a-16f3-4905-b3d9-783b0740eec7',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1115957553',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.419929] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.420105] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.420304] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.420455] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.420609] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.420818] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.420983] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.421168] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.421335] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.421509] env[62368]: DEBUG nova.virt.hardware [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.422278] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71c8d832-0141-42c1-a835-15153dae7461 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.427830] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 874.427830] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520184ae-2c4f-092a-9130-25fbfae80dc9" [ 874.427830] env[62368]: _type = "Task" [ 874.427830] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.435857] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520184ae-2c4f-092a-9130-25fbfae80dc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.583158] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.688950] env[62368]: DEBUG nova.network.neutron [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Updating instance_info_cache with network_info: [{"id": "cb4b1e4c-15e5-4641-b434-e05e31283e5c", "address": "fa:16:3e:22:8c:36", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb4b1e4c-15", "ovs_interfaceid": "cb4b1e4c-15e5-4641-b434-e05e31283e5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.939331] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520184ae-2c4f-092a-9130-25fbfae80dc9, 'name': SearchDatastore_Task, 'duration_secs': 0.009288} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.945178] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfiguring VM instance instance-0000003d to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 874.945493] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-841e65c1-6d6c-426e-a1d5-8afc9f47ebac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.966017] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 874.966017] env[62368]: value = "task-1198663" [ 874.966017] env[62368]: _type = "Task" [ 874.966017] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.974528] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198663, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.092075] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 875.092371] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.929s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.092731] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.746s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.094824] env[62368]: INFO nova.compute.claims [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.149588] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 875.192200] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "refresh_cache-69ddb565-6c79-44e5-a7d1-d339ab426fae" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.477675] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198663, 'name': ReconfigVM_Task, 'duration_secs': 0.236244} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.478011] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfigured VM instance instance-0000003d to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 875.478871] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc5476b-aa94-42d3-b590-69725b4af4f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.503134] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071/a0cea538-b162-4504-ac34-803a2d5a8071.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.503507] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ff87d75-e3a3-4944-a4c4-4224f8103896 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.524291] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 875.524291] env[62368]: value = "task-1198664" [ 875.524291] env[62368]: _type = "Task" [ 875.524291] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.533473] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198664, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.726319] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 875.726642] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1259bfc5-c04b-4233-b03d-bb1e579135b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.735525] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 875.735525] env[62368]: value = "task-1198665" [ 875.735525] env[62368]: _type = "Task" [ 875.735525] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.747025] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.035167] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198664, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.249059] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198665, 'name': PowerOffVM_Task, 'duration_secs': 0.19655} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.249698] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 876.250184] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318f4780-9d01-4ebd-a784-4d4c545a884b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.274682] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c7fe7a-f9a6-4327-945d-27da5ffb1940 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.302539] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 876.302866] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-affa2b01-4724-42d9-a66d-c41b87f34328 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.312393] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 876.312393] env[62368]: value = "task-1198666" [ 876.312393] env[62368]: _type = "Task" [ 876.312393] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.324668] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 876.324668] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.324867] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.325058] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.325243] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.326024] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56404fc4-b89b-4b5a-a2ae-de84e1dba635 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.338652] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.338856] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.339608] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93d23fc2-dda2-4bcb-905d-3ba871a83c3b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.345184] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 876.345184] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c49a70-ee02-6c22-3678-7a917c8e1722" [ 876.345184] env[62368]: _type = "Task" [ 876.345184] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.357564] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c49a70-ee02-6c22-3678-7a917c8e1722, 'name': SearchDatastore_Task, 'duration_secs': 0.009367} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.358352] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e06cac77-f8ac-4ad3-9c47-aa0438281a75 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.367162] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 876.367162] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5227abc6-8c40-5a92-6c3e-64c8abe431f9" [ 876.367162] env[62368]: _type = "Task" [ 876.367162] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.373653] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5227abc6-8c40-5a92-6c3e-64c8abe431f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.407794] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fed533c-4627-4d7f-a7c6-bad67d37708a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.415603] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aeb63a3-98b0-4e7a-8da5-33d036cfbf06 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.446634] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b1e624-15d5-463d-aa46-3171d6ac74a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.453963] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e06041-8b8c-4f76-9c69-053292b8fcba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.467193] env[62368]: DEBUG nova.compute.provider_tree [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.537019] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198664, 'name': ReconfigVM_Task, 'duration_secs': 0.72879} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.537705] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Reconfigured VM instance instance-0000003d to attach disk [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071/a0cea538-b162-4504-ac34-803a2d5a8071.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.538351] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846393fc-aa8d-4ec9-af06-d794fa8d61b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.558491] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b0312b-5fa2-44db-bf9b-66bb4c069baa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.576581] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79bb99f3-6585-4cb5-85d3-b248f99e3df3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.594584] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f3cc9c-a1c1-46bb-b280-8b9b0536d54e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.601603] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 876.601847] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c60ab093-5feb-46ee-ab2a-95a2bf74889a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.608370] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 876.608370] env[62368]: value = "task-1198667" [ 876.608370] env[62368]: _type = "Task" [ 876.608370] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.617447] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198667, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.875594] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5227abc6-8c40-5a92-6c3e-64c8abe431f9, 'name': SearchDatastore_Task, 'duration_secs': 0.009042} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.875903] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.876214] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. {{(pid=62368) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 876.876481] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31192dd6-f309-4135-81d7-e75bd12a0b94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.883985] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 876.883985] env[62368]: value = "task-1198668" [ 876.883985] env[62368]: _type = "Task" [ 876.883985] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.892374] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198668, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.970719] env[62368]: DEBUG nova.scheduler.client.report [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.121318] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198667, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.399221] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198668, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.476025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.383s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.476654] env[62368]: DEBUG nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.479731] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.120s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.479971] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.482913] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.958s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.483158] env[62368]: DEBUG nova.objects.instance [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 877.507499] env[62368]: INFO nova.scheduler.client.report [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted allocations for instance 417f000f-cf23-404d-877c-45990d1a7c77 [ 877.620043] env[62368]: DEBUG oslo_vmware.api [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198667, 'name': PowerOnVM_Task, 'duration_secs': 0.684585} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.621691] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 877.894820] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198668, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656233} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.895087] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. [ 877.896197] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b87b60c-bfa4-461b-8846-0d81a8bf4ad5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.921451] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.921735] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8f8df0a-875a-420e-bed1-a15ff0ec0d5f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.941880] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 877.941880] env[62368]: value = "task-1198669" [ 877.941880] env[62368]: _type = "Task" [ 877.941880] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.950268] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198669, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.992067] env[62368]: DEBUG nova.compute.utils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.993699] env[62368]: DEBUG nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.993908] env[62368]: DEBUG nova.network.neutron [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 878.016825] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1ee25a9f-1e24-48c0-9c70-604d7f2a63a3 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "417f000f-cf23-404d-877c-45990d1a7c77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.914s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.063337] env[62368]: DEBUG nova.policy [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f3afe492e7e47f28c3c8dad61f819d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81b69987192149ddb6dfcafc1f7770fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 878.371917] env[62368]: DEBUG nova.network.neutron [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Successfully created port: 52cba1c6-211a-412d-9265-34ce316a95a3 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.453157] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198669, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.495661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6c769dfa-8acc-4f13-b0e5-f1e35dc32018 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.499039] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.769s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.499278] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.501378] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.835s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.503031] env[62368]: INFO nova.compute.claims [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.506055] env[62368]: DEBUG nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.531320] env[62368]: INFO nova.scheduler.client.report [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Deleted allocations for instance 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5 [ 878.631855] env[62368]: INFO nova.compute.manager [None req-93002a9b-f34b-46de-b76c-0474b24bff15 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance to original state: 'active' [ 878.953783] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198669, 'name': ReconfigVM_Task, 'duration_secs': 0.812554} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.954662] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.955011] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4504ea60-f5ad-4020-9912-a5878ae4b66b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.980136] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-513d19b0-a20c-4729-a7bf-a62fd4fb785e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.996737] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 878.996737] env[62368]: value = "task-1198670" [ 878.996737] env[62368]: _type = "Task" [ 878.996737] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.006175] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198670, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.039665] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f3345bd3-b973-4c1b-a215-7bdc9c1ac042 tempest-ServersNegativeTestMultiTenantJSON-1190259321 tempest-ServersNegativeTestMultiTenantJSON-1190259321-project-member] Lock "044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.335s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.525111] env[62368]: DEBUG nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.527135] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198670, 'name': ReconfigVM_Task, 'duration_secs': 0.177196} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.529066] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 879.529066] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69607083-4551-4cb5-9143-233d54c85c8f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.539070] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 879.539070] env[62368]: value = "task-1198671" [ 879.539070] env[62368]: _type = "Task" [ 879.539070] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.546614] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.556211] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.556539] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.556820] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.557108] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.557311] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.557506] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.557891] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.558132] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.558353] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.558561] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.558781] env[62368]: DEBUG nova.virt.hardware [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.559704] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae194c96-366d-48f8-86f6-c7f1fec6ad1b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.568511] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "a0cea538-b162-4504-ac34-803a2d5a8071" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.568815] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "a0cea538-b162-4504-ac34-803a2d5a8071" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.569407] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.569407] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.569609] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "a0cea538-b162-4504-ac34-803a2d5a8071-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.572369] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9675884f-8c11-4455-beb8-9c7df60bcd00 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.577414] env[62368]: INFO nova.compute.manager [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Terminating instance [ 879.582533] env[62368]: DEBUG nova.compute.manager [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 879.582828] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 879.583729] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa9b996-9e49-4255-bc5e-7743bd392031 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.603502] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 879.603874] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28713dd4-3947-468c-9c0d-50fc519ad007 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.612067] env[62368]: DEBUG oslo_vmware.api [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 879.612067] env[62368]: value = "task-1198672" [ 879.612067] env[62368]: _type = "Task" [ 879.612067] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.621550] env[62368]: DEBUG oslo_vmware.api [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198672, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.819080] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e11649d-b261-435f-ad2f-c742bc89df69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.827014] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75493c27-bdde-40e9-850a-ce35e192acde {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.859867] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cb0e76-8bab-4716-a35e-2f1633ea7258 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.868796] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308f9076-75d6-49e3-b059-0a25a33506f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.885074] env[62368]: DEBUG nova.compute.provider_tree [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.040313] env[62368]: DEBUG nova.compute.manager [req-16f375e2-e45d-4ff5-9c3a-79daed5ff44e req-e3a1e836-5b36-41d8-830e-b67e809afe26 service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Received event network-vif-plugged-52cba1c6-211a-412d-9265-34ce316a95a3 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.040563] env[62368]: DEBUG oslo_concurrency.lockutils [req-16f375e2-e45d-4ff5-9c3a-79daed5ff44e req-e3a1e836-5b36-41d8-830e-b67e809afe26 service nova] Acquiring lock "55eecf3d-501c-490d-bddd-0211fd082841-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.040802] env[62368]: DEBUG oslo_concurrency.lockutils [req-16f375e2-e45d-4ff5-9c3a-79daed5ff44e req-e3a1e836-5b36-41d8-830e-b67e809afe26 service nova] Lock "55eecf3d-501c-490d-bddd-0211fd082841-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.040982] env[62368]: DEBUG oslo_concurrency.lockutils [req-16f375e2-e45d-4ff5-9c3a-79daed5ff44e req-e3a1e836-5b36-41d8-830e-b67e809afe26 service nova] Lock "55eecf3d-501c-490d-bddd-0211fd082841-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.041549] env[62368]: DEBUG nova.compute.manager [req-16f375e2-e45d-4ff5-9c3a-79daed5ff44e req-e3a1e836-5b36-41d8-830e-b67e809afe26 service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] No waiting events found dispatching network-vif-plugged-52cba1c6-211a-412d-9265-34ce316a95a3 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.041740] env[62368]: WARNING nova.compute.manager [req-16f375e2-e45d-4ff5-9c3a-79daed5ff44e req-e3a1e836-5b36-41d8-830e-b67e809afe26 service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Received unexpected event network-vif-plugged-52cba1c6-211a-412d-9265-34ce316a95a3 for instance with vm_state building and task_state spawning. [ 880.051991] env[62368]: DEBUG oslo_vmware.api [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198671, 'name': PowerOnVM_Task, 'duration_secs': 0.401397} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.052253] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 880.055039] env[62368]: DEBUG nova.compute.manager [None req-1d1399a8-c318-4c05-830d-680335349cf9 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.055868] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651e370e-ea57-4b32-ad9b-0c3ae3c50a57 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.122974] env[62368]: DEBUG oslo_vmware.api [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198672, 'name': PowerOffVM_Task, 'duration_secs': 0.232255} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.123290] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 880.123488] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 880.123803] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-055723ec-a9e2-470a-9caa-6cf522fc8471 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.201081] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 880.201340] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 880.201526] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Deleting the datastore file [datastore2] a0cea538-b162-4504-ac34-803a2d5a8071 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.201802] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4064e78b-7ef8-44ae-9aff-92f902d8e5f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.209121] env[62368]: DEBUG oslo_vmware.api [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for the task: (returnval){ [ 880.209121] env[62368]: value = "task-1198674" [ 880.209121] env[62368]: _type = "Task" [ 880.209121] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.218061] env[62368]: DEBUG oslo_vmware.api [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.237254] env[62368]: DEBUG nova.network.neutron [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Successfully updated port: 52cba1c6-211a-412d-9265-34ce316a95a3 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.388534] env[62368]: DEBUG nova.scheduler.client.report [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.720874] env[62368]: DEBUG oslo_vmware.api [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Task: {'id': task-1198674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158045} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.721411] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.721455] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 880.721719] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 880.722318] env[62368]: INFO nova.compute.manager [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Took 1.14 seconds to destroy the instance on the hypervisor. [ 880.722318] env[62368]: DEBUG oslo.service.loopingcall [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.722507] env[62368]: DEBUG nova.compute.manager [-] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.722748] env[62368]: DEBUG nova.network.neutron [-] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.744226] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "refresh_cache-55eecf3d-501c-490d-bddd-0211fd082841" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.744575] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "refresh_cache-55eecf3d-501c-490d-bddd-0211fd082841" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.748026] env[62368]: DEBUG nova.network.neutron [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.896205] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.896205] env[62368]: DEBUG nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 880.902423] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 15.773s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.285150] env[62368]: DEBUG nova.network.neutron [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 881.407697] env[62368]: DEBUG nova.compute.utils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.409725] env[62368]: INFO nova.compute.claims [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.415402] env[62368]: DEBUG nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 881.415402] env[62368]: DEBUG nova.network.neutron [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 881.490955] env[62368]: DEBUG nova.network.neutron [-] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.529066] env[62368]: DEBUG nova.policy [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0800ab273ca04fbf9396175b57eed6b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e821059910b4e32aab596c6f4d521d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.571685] env[62368]: DEBUG nova.network.neutron [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Updating instance_info_cache with network_info: [{"id": "52cba1c6-211a-412d-9265-34ce316a95a3", "address": "fa:16:3e:44:d5:b1", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52cba1c6-21", "ovs_interfaceid": "52cba1c6-211a-412d-9265-34ce316a95a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.860133] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.915724] env[62368]: DEBUG nova.network.neutron [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Successfully created port: 8be1ff35-0e69-42f1-a582-16bf1f496a2d {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.920047] env[62368]: INFO nova.compute.resource_tracker [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating resource usage from migration a4a55aca-f211-40de-8c6e-9835ad122ae2 [ 881.922598] env[62368]: DEBUG nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 881.994224] env[62368]: INFO nova.compute.manager [-] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Took 1.27 seconds to deallocate network for instance. [ 882.076297] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "refresh_cache-55eecf3d-501c-490d-bddd-0211fd082841" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.076958] env[62368]: DEBUG nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Instance network_info: |[{"id": "52cba1c6-211a-412d-9265-34ce316a95a3", "address": "fa:16:3e:44:d5:b1", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52cba1c6-21", "ovs_interfaceid": "52cba1c6-211a-412d-9265-34ce316a95a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 882.077308] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:d5:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52cba1c6-211a-412d-9265-34ce316a95a3', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.085385] env[62368]: DEBUG oslo.service.loopingcall [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.086027] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 882.086283] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75aec6f3-fbd7-41f9-8217-ffa9f4f92960 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.116915] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.116915] env[62368]: value = "task-1198675" [ 882.116915] env[62368]: _type = "Task" [ 882.116915] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.140930] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198675, 'name': CreateVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.213026] env[62368]: DEBUG nova.compute.manager [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Received event network-changed-52cba1c6-211a-412d-9265-34ce316a95a3 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.213026] env[62368]: DEBUG nova.compute.manager [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Refreshing instance network info cache due to event network-changed-52cba1c6-211a-412d-9265-34ce316a95a3. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 882.213127] env[62368]: DEBUG oslo_concurrency.lockutils [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] Acquiring lock "refresh_cache-55eecf3d-501c-490d-bddd-0211fd082841" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.213219] env[62368]: DEBUG oslo_concurrency.lockutils [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] Acquired lock "refresh_cache-55eecf3d-501c-490d-bddd-0211fd082841" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.213384] env[62368]: DEBUG nova.network.neutron [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Refreshing network info cache for port 52cba1c6-211a-412d-9265-34ce316a95a3 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 882.275327] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8030649-bbf2-46f2-af08-c04b3d0fdd95 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.283686] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2b2a34-8864-4e8a-a017-d435ce1dd675 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.322624] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11578bf-a2a4-4dd3-b04b-7f011c11ca1b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.331095] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748ee4b6-1610-4b85-bbdc-47d3827bfab4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.347501] env[62368]: DEBUG nova.compute.provider_tree [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.501598] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.609310] env[62368]: INFO nova.compute.manager [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Rescuing [ 882.609523] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.609628] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.609957] env[62368]: DEBUG nova.network.neutron [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.627843] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198675, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.851588] env[62368]: DEBUG nova.scheduler.client.report [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.860287] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 882.932251] env[62368]: DEBUG nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 882.957839] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.958146] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.958321] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.958512] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.958658] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.958806] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.959031] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.959200] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.959367] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.959528] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.960094] env[62368]: DEBUG nova.virt.hardware [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.960603] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c8c4fe-6687-4f4e-86a4-43071022ed64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.971349] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435141d9-a2d8-4356-a09d-6235e395f6e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.130247] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198675, 'name': CreateVM_Task, 'duration_secs': 0.530408} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.130435] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 883.131196] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.131374] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.131735] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.134053] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86495d86-c1e4-4b13-99b9-54ad316151f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.138380] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 883.138380] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f4d5c8-e4cd-82ee-be72-33f982065cd9" [ 883.138380] env[62368]: _type = "Task" [ 883.138380] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.149081] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f4d5c8-e4cd-82ee-be72-33f982065cd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.178454] env[62368]: DEBUG nova.network.neutron [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Updated VIF entry in instance network info cache for port 52cba1c6-211a-412d-9265-34ce316a95a3. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 883.178846] env[62368]: DEBUG nova.network.neutron [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Updating instance_info_cache with network_info: [{"id": "52cba1c6-211a-412d-9265-34ce316a95a3", "address": "fa:16:3e:44:d5:b1", "network": {"id": "3202b12c-8f9b-4348-9951-4a0ae495760a", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1570691692-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81b69987192149ddb6dfcafc1f7770fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52cba1c6-21", "ovs_interfaceid": "52cba1c6-211a-412d-9265-34ce316a95a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.358087] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.456s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.358508] env[62368]: INFO nova.compute.manager [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Migrating [ 883.365838] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.472s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.371028] env[62368]: INFO nova.compute.claims [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.435455] env[62368]: DEBUG nova.network.neutron [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Updating instance_info_cache with network_info: [{"id": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "address": "fa:16:3e:fe:ae:1b", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b605458-f2", "ovs_interfaceid": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.611062] env[62368]: DEBUG nova.network.neutron [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Successfully updated port: 8be1ff35-0e69-42f1-a582-16bf1f496a2d {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.651456] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f4d5c8-e4cd-82ee-be72-33f982065cd9, 'name': SearchDatastore_Task, 'duration_secs': 0.027666} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.651784] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.652042] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.652335] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.652505] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.652697] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.652970] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec77dd3e-7531-48d9-b1c6-6805fe248498 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.662685] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.662890] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 883.663638] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16c6bcf5-2eaf-43af-a91d-ef45c082467c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.669899] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 883.669899] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520798b4-c58f-be8c-f36f-8f77277bc8e7" [ 883.669899] env[62368]: _type = "Task" [ 883.669899] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.678882] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520798b4-c58f-be8c-f36f-8f77277bc8e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.681551] env[62368]: DEBUG oslo_concurrency.lockutils [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] Releasing lock "refresh_cache-55eecf3d-501c-490d-bddd-0211fd082841" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.683348] env[62368]: DEBUG nova.compute.manager [req-4f91e059-093d-4ae1-b0c3-a2e9ba59fbf8 req-3f94c74a-1d18-43c5-a380-aafacbfa96dd service nova] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Received event network-vif-deleted-444a47e3-2f56-4895-8a8b-7e5605dbfaf1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.882221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.882221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.882221] env[62368]: DEBUG nova.network.neutron [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 883.938716] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.963678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.963902] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.113820] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.113985] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.114157] env[62368]: DEBUG nova.network.neutron [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 884.181714] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520798b4-c58f-be8c-f36f-8f77277bc8e7, 'name': SearchDatastore_Task, 'duration_secs': 0.017184} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.182618] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49877e0e-f06b-41d9-9111-55f446bd6079 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.188925] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 884.188925] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52bb0afc-6e7d-638f-ec7c-2d518a343c00" [ 884.188925] env[62368]: _type = "Task" [ 884.188925] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.197468] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52bb0afc-6e7d-638f-ec7c-2d518a343c00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.244544] env[62368]: DEBUG nova.compute.manager [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Received event network-vif-plugged-8be1ff35-0e69-42f1-a582-16bf1f496a2d {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.246182] env[62368]: DEBUG oslo_concurrency.lockutils [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] Acquiring lock "5af32858-dc9a-4380-827a-daf384c29bed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.246182] env[62368]: DEBUG oslo_concurrency.lockutils [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] Lock "5af32858-dc9a-4380-827a-daf384c29bed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.246182] env[62368]: DEBUG oslo_concurrency.lockutils [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] Lock "5af32858-dc9a-4380-827a-daf384c29bed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.246182] env[62368]: DEBUG nova.compute.manager [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] No waiting events found dispatching network-vif-plugged-8be1ff35-0e69-42f1-a582-16bf1f496a2d {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.246182] env[62368]: WARNING nova.compute.manager [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Received unexpected event network-vif-plugged-8be1ff35-0e69-42f1-a582-16bf1f496a2d for instance with vm_state building and task_state spawning. [ 884.246182] env[62368]: DEBUG nova.compute.manager [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Received event network-changed-8be1ff35-0e69-42f1-a582-16bf1f496a2d {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.246700] env[62368]: DEBUG nova.compute.manager [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Refreshing instance network info cache due to event network-changed-8be1ff35-0e69-42f1-a582-16bf1f496a2d. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 884.246700] env[62368]: DEBUG oslo_concurrency.lockutils [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] Acquiring lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.468092] env[62368]: DEBUG nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.479741] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 884.479997] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87d28c57-3d07-405a-8a17-5dbeb35c08b7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.493358] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 884.493358] env[62368]: value = "task-1198676" [ 884.493358] env[62368]: _type = "Task" [ 884.493358] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.504958] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.663507] env[62368]: DEBUG nova.network.neutron [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 884.693459] env[62368]: DEBUG nova.network.neutron [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.697374] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76743ffa-0271-485c-b599-596f1f341c2c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.704914] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52bb0afc-6e7d-638f-ec7c-2d518a343c00, 'name': SearchDatastore_Task, 'duration_secs': 0.010822} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.705920] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.706258] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 55eecf3d-501c-490d-bddd-0211fd082841/55eecf3d-501c-490d-bddd-0211fd082841.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 884.707326] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e80dd77a-c72f-4984-bce7-2ba964be1422 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.712886] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60d7901-c3d7-48ab-b77e-7813c7eb8cc3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.717956] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 884.717956] env[62368]: value = "task-1198677" [ 884.717956] env[62368]: _type = "Task" [ 884.717956] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.751533] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6bf444b-838d-4c6e-86e4-44c2f0840874 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.757829] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198677, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.764041] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b4abf0-a48b-4f57-b8b3-3e394920e53a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.779169] env[62368]: DEBUG nova.compute.provider_tree [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.857987] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 884.859694] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 884.859914] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 884.885097] env[62368]: DEBUG nova.network.neutron [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance_info_cache with network_info: [{"id": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "address": "fa:16:3e:e1:de:4b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be1ff35-0e", "ovs_interfaceid": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.999523] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.005997] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198676, 'name': PowerOffVM_Task, 'duration_secs': 0.204768} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.006327] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 885.007158] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750fe892-995b-421f-9faf-d9e7aa1a2f59 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.028276] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee25d59-bffb-4a5e-a08f-5506476e5f8d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.065839] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 885.066236] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46fddc92-a615-442f-92ba-df743564ee7b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.074861] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 885.074861] env[62368]: value = "task-1198678" [ 885.074861] env[62368]: _type = "Task" [ 885.074861] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.085170] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 885.085459] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.085661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.085815] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.086015] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.086357] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fc6788e-6563-46a5-9156-24843d71213a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.100182] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.100393] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 885.101395] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c25f52e3-7046-4cda-8bd4-ba9ac2e2abfe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.108757] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 885.108757] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522c5139-2f04-4ce7-ec76-80b23e1fc4bf" [ 885.108757] env[62368]: _type = "Task" [ 885.108757] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.118210] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522c5139-2f04-4ce7-ec76-80b23e1fc4bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.206029] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.229698] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198677, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465824} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.229868] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 55eecf3d-501c-490d-bddd-0211fd082841/55eecf3d-501c-490d-bddd-0211fd082841.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 885.230022] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.230282] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1402eac-e8f2-43bd-9e9c-6d13e0770075 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.237040] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 885.237040] env[62368]: value = "task-1198679" [ 885.237040] env[62368]: _type = "Task" [ 885.237040] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.244958] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.284256] env[62368]: DEBUG nova.scheduler.client.report [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.387544] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.388379] env[62368]: DEBUG nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Instance network_info: |[{"id": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "address": "fa:16:3e:e1:de:4b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be1ff35-0e", "ovs_interfaceid": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.388379] env[62368]: DEBUG oslo_concurrency.lockutils [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] Acquired lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.388583] env[62368]: DEBUG nova.network.neutron [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Refreshing network info cache for port 8be1ff35-0e69-42f1-a582-16bf1f496a2d {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 885.390370] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:de:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8be1ff35-0e69-42f1-a582-16bf1f496a2d', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.397677] env[62368]: DEBUG oslo.service.loopingcall [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.398592] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 885.398830] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2654f11-881d-4862-8ff9-0e06990cb4c5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.419193] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.419193] env[62368]: value = "task-1198680" [ 885.419193] env[62368]: _type = "Task" [ 885.419193] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.427615] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198680, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.621653] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522c5139-2f04-4ce7-ec76-80b23e1fc4bf, 'name': SearchDatastore_Task, 'duration_secs': 0.023475} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.622466] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0e1026a-fd40-4370-8368-ed9b5539b9bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.628381] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 885.628381] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5252e6c5-62bb-c2c7-940f-55ee1af5fa01" [ 885.628381] env[62368]: _type = "Task" [ 885.628381] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.635812] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5252e6c5-62bb-c2c7-940f-55ee1af5fa01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.747993] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099689} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.747993] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.748827] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea45dbd-24ab-4bd2-b30c-5132db06f743 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.773206] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 55eecf3d-501c-490d-bddd-0211fd082841/55eecf3d-501c-490d-bddd-0211fd082841.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.773825] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-944d34ae-b0bd-4bab-9e7a-7e588eecd69f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.789379] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.789882] env[62368]: DEBUG nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.792843] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.792s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.792957] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.795024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.568s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.795225] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.796923] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.160s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.797196] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.799010] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.314s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.799220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.801088] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.260s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.801279] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.802827] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.250s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.803120] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.804596] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.474s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.806977] env[62368]: INFO nova.compute.claims [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.816803] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 885.816803] env[62368]: value = "task-1198681" [ 885.816803] env[62368]: _type = "Task" [ 885.816803] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.826702] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198681, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.831425] env[62368]: INFO nova.scheduler.client.report [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleted allocations for instance e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f [ 885.841010] env[62368]: INFO nova.scheduler.client.report [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted allocations for instance 58efc1cf-2469-41be-bf99-fc7b8c72113c [ 885.855741] env[62368]: INFO nova.scheduler.client.report [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted allocations for instance 5554732b-34a1-41bd-9d0d-9bd39d62bec3 [ 885.864590] env[62368]: INFO nova.scheduler.client.report [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted allocations for instance 5ed9cee5-8cc8-4b70-96a9-737019363638 [ 885.867854] env[62368]: INFO nova.scheduler.client.report [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleted allocations for instance f397e102-bfbd-4753-b1a4-4c5e9126b2c6 [ 885.876920] env[62368]: INFO nova.scheduler.client.report [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Deleted allocations for instance cda0cf42-11ae-4da3-b838-5bac4bc2f4d1 [ 885.879759] env[62368]: WARNING oslo_messaging._drivers.amqpdriver [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 885.931495] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198680, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.143749] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5252e6c5-62bb-c2c7-940f-55ee1af5fa01, 'name': SearchDatastore_Task, 'duration_secs': 0.012773} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.144244] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.144386] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b87094da-6258-469e-ab37-5557955ad3a7/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. {{(pid=62368) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 886.144657] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd3fd98b-8516-472d-93ea-f2bc5394a3e0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.168272] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 886.168272] env[62368]: value = "task-1198682" [ 886.168272] env[62368]: _type = "Task" [ 886.168272] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.177039] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198682, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.191944] env[62368]: DEBUG nova.network.neutron [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updated VIF entry in instance network info cache for port 8be1ff35-0e69-42f1-a582-16bf1f496a2d. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 886.192379] env[62368]: DEBUG nova.network.neutron [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance_info_cache with network_info: [{"id": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "address": "fa:16:3e:e1:de:4b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be1ff35-0e", "ovs_interfaceid": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.311939] env[62368]: DEBUG nova.compute.utils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 886.315687] env[62368]: DEBUG nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 886.315886] env[62368]: DEBUG nova.network.neutron [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 886.328494] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198681, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.339981] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9167b4f4-6e0d-4c47-8c49-d0b84d5095d0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.886s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.353670] env[62368]: DEBUG oslo_concurrency.lockutils [None req-628ff8ca-990f-4d65-9ea6-5af31610feb3 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "58efc1cf-2469-41be-bf99-fc7b8c72113c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.281s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.363185] env[62368]: DEBUG nova.policy [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6972fdf35fc4fe9afa65cdfd0d46568', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb560320635f493fbe4308af1623e8ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.367507] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e022c4c1-9ea9-4281-bcec-4805fb5f23ae tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "5554732b-34a1-41bd-9d0d-9bd39d62bec3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.177s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.371568] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8bc389e9-d1e4-4e89-86db-45232fe02432 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "5ed9cee5-8cc8-4b70-96a9-737019363638" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.551s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.375571] env[62368]: DEBUG oslo_concurrency.lockutils [None req-11b62375-c064-4311-b49f-d331abc43759 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "f397e102-bfbd-4753-b1a4-4c5e9126b2c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.805s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.383447] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1f0d3795-275d-4cc2-a698-3a72f0448cc0 tempest-ListServersNegativeTestJSON-1586818397 tempest-ListServersNegativeTestJSON-1586818397-project-member] Lock "cda0cf42-11ae-4da3-b838-5bac4bc2f4d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.007s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.435978] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198680, 'name': CreateVM_Task, 'duration_secs': 0.859586} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.436307] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 886.437380] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.437581] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.438099] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 886.438506] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3dff876-0b95-4d24-a458-dc718aa2bd26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.448191] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 886.448191] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52d0d974-64bd-8489-ea20-a4a072894ffe" [ 886.448191] env[62368]: _type = "Task" [ 886.448191] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.460866] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d0d974-64bd-8489-ea20-a4a072894ffe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.681334] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198682, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.695427] env[62368]: DEBUG oslo_concurrency.lockutils [req-8f09bfc3-8986-4705-a659-16819a8d09be req-93b85079-2235-4c5c-bad0-2cb1a1cf07a0 service nova] Releasing lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.723699] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e781f3-1ba9-46d5-8ec3-354e95bd0e50 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.751323] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance '13765305-2e55-4ee8-9a6f-4ae5ee724367' progress to 0 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 886.761361] env[62368]: DEBUG nova.network.neutron [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Successfully created port: bd3ff77d-c1de-422a-81ca-03f3aa55bd9a {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.821641] env[62368]: DEBUG nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.834670] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198681, 'name': ReconfigVM_Task, 'duration_secs': 0.721641} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.835146] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 55eecf3d-501c-490d-bddd-0211fd082841/55eecf3d-501c-490d-bddd-0211fd082841.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.835952] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-25acd793-6e17-4ac1-9113-3c32aa5417b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.843489] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 886.843489] env[62368]: value = "task-1198683" [ 886.843489] env[62368]: _type = "Task" [ 886.843489] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.855679] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198683, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.964017] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d0d974-64bd-8489-ea20-a4a072894ffe, 'name': SearchDatastore_Task, 'duration_secs': 0.017873} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.964394] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.964655] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.964960] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.965153] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.965377] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.965708] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-693482c3-f2c9-4592-b872-aacc3c0a83c3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.979620] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.979824] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 886.983168] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88c1a3dc-3c11-4b57-bfcb-289e45963481 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.990618] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 886.990618] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5284d576-86f5-694b-21c3-b58752e79cb5" [ 886.990618] env[62368]: _type = "Task" [ 886.990618] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.999826] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5284d576-86f5-694b-21c3-b58752e79cb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.056965] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6df2845-2ee4-4d74-882c-9ce612784108 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.065444] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a42cbd-dacb-4bbe-8a06-1b6ff1c01284 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.095478] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70af80d7-70cc-4417-be15-0abc9ea220d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.103977] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bc4e98-e159-43fe-b8b5-4e6b3db9bbc8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.118635] env[62368]: DEBUG nova.compute.provider_tree [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.179223] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198682, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606124} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.179501] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b87094da-6258-469e-ab37-5557955ad3a7/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. [ 887.180321] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbe2dc1-288e-42cf-82e5-dd836987baca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.210924] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] b87094da-6258-469e-ab37-5557955ad3a7/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.210924] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d129caa2-9927-4f70-ac11-b2f0a5e659f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.232695] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 887.232695] env[62368]: value = "task-1198684" [ 887.232695] env[62368]: _type = "Task" [ 887.232695] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.242224] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198684, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.259020] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 887.259020] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa2f91fe-de2e-4228-9b20-d293447369f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.271111] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 887.271111] env[62368]: value = "task-1198685" [ 887.271111] env[62368]: _type = "Task" [ 887.271111] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.281526] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 887.281752] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance '13765305-2e55-4ee8-9a6f-4ae5ee724367' progress to 17 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 887.360427] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198683, 'name': Rename_Task, 'duration_secs': 0.19206} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.360427] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 887.360427] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85bba566-ce01-4052-a151-3a5f299d66de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.369067] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 887.369067] env[62368]: value = "task-1198686" [ 887.369067] env[62368]: _type = "Task" [ 887.369067] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.381041] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198686, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.502839] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5284d576-86f5-694b-21c3-b58752e79cb5, 'name': SearchDatastore_Task, 'duration_secs': 0.011963} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.504853] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70c037d5-601a-4589-8417-450e30e645d9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.509936] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 887.509936] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524a689b-9f97-c3f4-391d-d97a33d48dfc" [ 887.509936] env[62368]: _type = "Task" [ 887.509936] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.519505] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524a689b-9f97-c3f4-391d-d97a33d48dfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.621503] env[62368]: DEBUG nova.scheduler.client.report [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.748241] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198684, 'name': ReconfigVM_Task, 'duration_secs': 0.424014} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.748693] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Reconfigured VM instance instance-0000004e to attach disk [datastore1] b87094da-6258-469e-ab37-5557955ad3a7/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.749844] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe5cb33-6de1-4831-acba-98044fb5c4fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.779807] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-606fee54-c4d8-4857-8c47-d11b5ced9f1c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.793870] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.794175] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.794295] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.794722] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.794935] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.795113] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.795332] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.795499] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.795725] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.795905] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.796104] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.801499] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b19257f2-e0e8-4314-b67d-588d128ad8dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.823807] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 887.823807] env[62368]: value = "task-1198688" [ 887.823807] env[62368]: _type = "Task" [ 887.823807] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.825577] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 887.825577] env[62368]: value = "task-1198687" [ 887.825577] env[62368]: _type = "Task" [ 887.825577] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.832920] env[62368]: DEBUG nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.846026] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198688, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.850295] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198687, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.865030] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.865840] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.865840] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.865840] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.865840] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.866354] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.867496] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.867496] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.867496] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.867496] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.867699] env[62368]: DEBUG nova.virt.hardware [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.868613] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed27d96-e4de-47c0-b57e-75c0e1041774 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.894448] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198686, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.897906] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f94166-49ef-47ab-9999-37a2a3290ae1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.915399] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.915559] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquired lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.915711] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Forcefully refreshing network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 888.022670] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524a689b-9f97-c3f4-391d-d97a33d48dfc, 'name': SearchDatastore_Task, 'duration_secs': 0.011946} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.023035] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.023400] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 5af32858-dc9a-4380-827a-daf384c29bed/5af32858-dc9a-4380-827a-daf384c29bed.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 888.023690] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46d812c2-7f0f-4cc7-a4f8-022aecb29415 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.031892] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 888.031892] env[62368]: value = "task-1198689" [ 888.031892] env[62368]: _type = "Task" [ 888.031892] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.044615] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198689, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.127714] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.128418] env[62368]: DEBUG nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 888.133052] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.884s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.133052] env[62368]: DEBUG nova.objects.instance [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lazy-loading 'resources' on Instance uuid 0cb0097b-2908-4107-bb57-0241915ccff7 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.288318] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "ad21bf3a-068e-44a7-b58f-2978a631b972" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.290556] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "ad21bf3a-068e-44a7-b58f-2978a631b972" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.340593] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198688, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.347137] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198687, 'name': ReconfigVM_Task, 'duration_secs': 0.325403} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.347499] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance '13765305-2e55-4ee8-9a6f-4ae5ee724367' progress to 33 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 888.387693] env[62368]: DEBUG oslo_vmware.api [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198686, 'name': PowerOnVM_Task, 'duration_secs': 0.544793} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.388292] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 888.388527] env[62368]: INFO nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Took 8.86 seconds to spawn the instance on the hypervisor. [ 888.388718] env[62368]: DEBUG nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.389571] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914f5b8a-fbc6-4050-afa6-7120aa76750a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.448650] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 888.547533] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198689, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50875} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.548722] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 5af32858-dc9a-4380-827a-daf384c29bed/5af32858-dc9a-4380-827a-daf384c29bed.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 888.548722] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.549287] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-97436da0-f7f2-427a-856f-c9d27feb291f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.564029] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 888.564029] env[62368]: value = "task-1198690" [ 888.564029] env[62368]: _type = "Task" [ 888.564029] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.571944] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.638007] env[62368]: DEBUG nova.compute.utils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.643199] env[62368]: DEBUG nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 888.643422] env[62368]: DEBUG nova.network.neutron [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 888.670825] env[62368]: DEBUG nova.compute.manager [req-d900a940-8b40-4995-aad8-48865f9a7cf2 req-8a45816a-5875-4e58-a740-292e71fd08da service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Received event network-vif-plugged-bd3ff77d-c1de-422a-81ca-03f3aa55bd9a {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.670877] env[62368]: DEBUG oslo_concurrency.lockutils [req-d900a940-8b40-4995-aad8-48865f9a7cf2 req-8a45816a-5875-4e58-a740-292e71fd08da service nova] Acquiring lock "207de202-ca50-4811-84c5-a0d4454d29f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.671062] env[62368]: DEBUG oslo_concurrency.lockutils [req-d900a940-8b40-4995-aad8-48865f9a7cf2 req-8a45816a-5875-4e58-a740-292e71fd08da service nova] Lock "207de202-ca50-4811-84c5-a0d4454d29f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.671250] env[62368]: DEBUG oslo_concurrency.lockutils [req-d900a940-8b40-4995-aad8-48865f9a7cf2 req-8a45816a-5875-4e58-a740-292e71fd08da service nova] Lock "207de202-ca50-4811-84c5-a0d4454d29f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.671427] env[62368]: DEBUG nova.compute.manager [req-d900a940-8b40-4995-aad8-48865f9a7cf2 req-8a45816a-5875-4e58-a740-292e71fd08da service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] No waiting events found dispatching network-vif-plugged-bd3ff77d-c1de-422a-81ca-03f3aa55bd9a {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.671599] env[62368]: WARNING nova.compute.manager [req-d900a940-8b40-4995-aad8-48865f9a7cf2 req-8a45816a-5875-4e58-a740-292e71fd08da service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Received unexpected event network-vif-plugged-bd3ff77d-c1de-422a-81ca-03f3aa55bd9a for instance with vm_state building and task_state spawning. [ 888.693138] env[62368]: DEBUG nova.policy [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68c2209271e043b698fd3794955f40bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcecae6c09124e9fa4dcd5d63833b0bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.793947] env[62368]: DEBUG nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.840538] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198688, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.857140] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.857563] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.857779] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.858236] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.858447] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.858648] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.858896] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.859076] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.859303] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.859426] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.860059] env[62368]: DEBUG nova.virt.hardware [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.866413] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Reconfiguring VM instance instance-00000032 to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 888.869781] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e888d230-9291-49f0-93da-467a947773c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.895253] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 888.895253] env[62368]: value = "task-1198691" [ 888.895253] env[62368]: _type = "Task" [ 888.895253] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.910458] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198691, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.915281] env[62368]: INFO nova.compute.manager [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Took 28.59 seconds to build instance. [ 888.944109] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d01000-f274-4357-ae13-ffa14aa94806 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.956609] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4155e597-a0c1-46a3-b962-3d804c3f29cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.994124] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6dd21c-31ab-4d71-8445-b349cbed6cb4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.003944] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f6c51b-241a-42e0-900b-b94fa4eb2934 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.020173] env[62368]: DEBUG nova.compute.provider_tree [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.022893] env[62368]: DEBUG nova.network.neutron [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Successfully updated port: bd3ff77d-c1de-422a-81ca-03f3aa55bd9a {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.048529] env[62368]: DEBUG nova.compute.manager [req-d84d5a39-81df-478a-8217-f6be107085ff req-dfd5ba30-ec65-4dd3-88dd-c7414335617a service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Received event network-changed-bd3ff77d-c1de-422a-81ca-03f3aa55bd9a {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.048770] env[62368]: DEBUG nova.compute.manager [req-d84d5a39-81df-478a-8217-f6be107085ff req-dfd5ba30-ec65-4dd3-88dd-c7414335617a service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Refreshing instance network info cache due to event network-changed-bd3ff77d-c1de-422a-81ca-03f3aa55bd9a. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.048996] env[62368]: DEBUG oslo_concurrency.lockutils [req-d84d5a39-81df-478a-8217-f6be107085ff req-dfd5ba30-ec65-4dd3-88dd-c7414335617a service nova] Acquiring lock "refresh_cache-207de202-ca50-4811-84c5-a0d4454d29f7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.049164] env[62368]: DEBUG oslo_concurrency.lockutils [req-d84d5a39-81df-478a-8217-f6be107085ff req-dfd5ba30-ec65-4dd3-88dd-c7414335617a service nova] Acquired lock "refresh_cache-207de202-ca50-4811-84c5-a0d4454d29f7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.049325] env[62368]: DEBUG nova.network.neutron [req-d84d5a39-81df-478a-8217-f6be107085ff req-dfd5ba30-ec65-4dd3-88dd-c7414335617a service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Refreshing network info cache for port bd3ff77d-c1de-422a-81ca-03f3aa55bd9a {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 889.073724] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197359} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.073990] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.074761] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f843e40-5b65-4f2f-b9d0-77463e783b3f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.100349] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 5af32858-dc9a-4380-827a-daf384c29bed/5af32858-dc9a-4380-827a-daf384c29bed.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.101013] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10c99520-4911-47ca-89dd-3ef5cd53dc2e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.123472] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 889.123472] env[62368]: value = "task-1198692" [ 889.123472] env[62368]: _type = "Task" [ 889.123472] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.129786] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.135766] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198692, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.140864] env[62368]: DEBUG nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 889.283212] env[62368]: DEBUG nova.network.neutron [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Successfully created port: c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 889.321802] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.338816] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198688, 'name': ReconfigVM_Task, 'duration_secs': 1.420609} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.339474] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 889.339474] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8f1ef74-aad1-44f6-9a5b-0e8aa4259823 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.347351] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 889.347351] env[62368]: value = "task-1198693" [ 889.347351] env[62368]: _type = "Task" [ 889.347351] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.358701] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198693, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.381255] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.381255] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.411580] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198691, 'name': ReconfigVM_Task, 'duration_secs': 0.455673} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.412134] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Reconfigured VM instance instance-00000032 to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 889.414020] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74dea9e1-6e95-4752-8245-9e5ba5622191 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.416578] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1088b991-fd16-484f-ab85-fd1627fb6159 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "55eecf3d-501c-490d-bddd-0211fd082841" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.100s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.440770] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 13765305-2e55-4ee8-9a6f-4ae5ee724367/13765305-2e55-4ee8-9a6f-4ae5ee724367.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.441289] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc9e43f8-86ae-46ce-9855-bf99a04642db {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.462179] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 889.462179] env[62368]: value = "task-1198694" [ 889.462179] env[62368]: _type = "Task" [ 889.462179] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.471889] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198694, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.526026] env[62368]: DEBUG nova.scheduler.client.report [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.527983] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "refresh_cache-207de202-ca50-4811-84c5-a0d4454d29f7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.601900] env[62368]: DEBUG nova.network.neutron [req-d84d5a39-81df-478a-8217-f6be107085ff req-dfd5ba30-ec65-4dd3-88dd-c7414335617a service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.633721] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198692, 'name': ReconfigVM_Task, 'duration_secs': 0.431814} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.634383] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 5af32858-dc9a-4380-827a-daf384c29bed/5af32858-dc9a-4380-827a-daf384c29bed.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.634586] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ce07c30-e0ab-45e8-aee7-a5ff4c604606 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.636413] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Releasing lock "refresh_cache-a0cea538-b162-4504-ac34-803a2d5a8071" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.636580] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Updated the network info_cache for instance {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 889.636761] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 889.637350] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 889.638183] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 889.638183] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 889.638183] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 889.638183] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 889.642258] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 889.642258] env[62368]: value = "task-1198695" [ 889.642258] env[62368]: _type = "Task" [ 889.642258] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.656257] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198695, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.771615] env[62368]: DEBUG nova.network.neutron [req-d84d5a39-81df-478a-8217-f6be107085ff req-dfd5ba30-ec65-4dd3-88dd-c7414335617a service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.857055] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198693, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.884594] env[62368]: DEBUG nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.978319] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198694, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.029853] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.032557] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.531s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.032557] env[62368]: DEBUG nova.objects.instance [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lazy-loading 'resources' on Instance uuid a0cea538-b162-4504-ac34-803a2d5a8071 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.054952] env[62368]: INFO nova.scheduler.client.report [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Deleted allocations for instance 0cb0097b-2908-4107-bb57-0241915ccff7 [ 890.141345] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.151518] env[62368]: DEBUG nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 890.169044] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198695, 'name': Rename_Task, 'duration_secs': 0.181822} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.171392] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 890.172213] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-275831d3-e1af-43ca-ae22-8265c7155467 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.178498] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.178793] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.179032] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.179208] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.179404] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.179579] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.179848] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.180086] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.180297] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.180512] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.180719] env[62368]: DEBUG nova.virt.hardware [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.181817] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5939901-eeba-49f0-a1c8-87de5b15c69a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.188977] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 890.188977] env[62368]: value = "task-1198696" [ 890.188977] env[62368]: _type = "Task" [ 890.188977] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.197035] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405f50ad-3650-483f-826f-eddf4970b867 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.204598] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.276357] env[62368]: DEBUG oslo_concurrency.lockutils [req-d84d5a39-81df-478a-8217-f6be107085ff req-dfd5ba30-ec65-4dd3-88dd-c7414335617a service nova] Releasing lock "refresh_cache-207de202-ca50-4811-84c5-a0d4454d29f7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.276768] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "refresh_cache-207de202-ca50-4811-84c5-a0d4454d29f7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.276938] env[62368]: DEBUG nova.network.neutron [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 890.359298] env[62368]: DEBUG oslo_vmware.api [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198693, 'name': PowerOnVM_Task, 'duration_secs': 0.905929} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.359641] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 890.363134] env[62368]: DEBUG nova.compute.manager [None req-5fff8936-cbcb-455a-8a8d-31919ead61cb tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.366342] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27101e57-6880-44a6-8395-3090ff565196 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.406824] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.473279] env[62368]: DEBUG oslo_vmware.api [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198694, 'name': ReconfigVM_Task, 'duration_secs': 0.809282} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.473593] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 13765305-2e55-4ee8-9a6f-4ae5ee724367/13765305-2e55-4ee8-9a6f-4ae5ee724367.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.473889] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance '13765305-2e55-4ee8-9a6f-4ae5ee724367' progress to 50 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 890.565028] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34576dfc-d317-449f-8cbb-7c09f9854fd1 tempest-InstanceActionsTestJSON-1852436585 tempest-InstanceActionsTestJSON-1852436585-project-member] Lock "0cb0097b-2908-4107-bb57-0241915ccff7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.058s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.701975] env[62368]: DEBUG oslo_vmware.api [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198696, 'name': PowerOnVM_Task, 'duration_secs': 0.504468} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.702278] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 890.702483] env[62368]: INFO nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Took 7.77 seconds to spawn the instance on the hypervisor. [ 890.702739] env[62368]: DEBUG nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.703549] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1598e3a-5797-47de-b3a6-87eca8f9d937 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.771148] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fee8be-a432-4de6-8dbe-68467fe4837b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.781976] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fb231d-9a51-4471-84e7-1f64de2a9f92 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.815721] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c2603e-c638-4330-9981-2206b4c75c41 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.824501] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bc01f4-b693-4ce6-a3b3-22a82adf2d13 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.840130] env[62368]: DEBUG nova.compute.provider_tree [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.842449] env[62368]: DEBUG nova.network.neutron [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 890.870850] env[62368]: DEBUG nova.compute.manager [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.871781] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea26b2cb-1495-452e-9f45-b7285f0ae5c1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.984691] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e74159-0966-4649-b35b-6211fe0e2df9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.019717] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49c5c85-1ee5-4fad-a7f3-6045833e6ce8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.050999] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance '13765305-2e55-4ee8-9a6f-4ae5ee724367' progress to 67 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 891.168937] env[62368]: DEBUG nova.compute.manager [req-18890692-ee3c-4029-91aa-5aefd4bf2c1a req-82edf882-b0db-40b7-80e8-31f89c691892 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-vif-plugged-c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.169209] env[62368]: DEBUG oslo_concurrency.lockutils [req-18890692-ee3c-4029-91aa-5aefd4bf2c1a req-82edf882-b0db-40b7-80e8-31f89c691892 service nova] Acquiring lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.169471] env[62368]: DEBUG oslo_concurrency.lockutils [req-18890692-ee3c-4029-91aa-5aefd4bf2c1a req-82edf882-b0db-40b7-80e8-31f89c691892 service nova] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.169690] env[62368]: DEBUG oslo_concurrency.lockutils [req-18890692-ee3c-4029-91aa-5aefd4bf2c1a req-82edf882-b0db-40b7-80e8-31f89c691892 service nova] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.170037] env[62368]: DEBUG nova.compute.manager [req-18890692-ee3c-4029-91aa-5aefd4bf2c1a req-82edf882-b0db-40b7-80e8-31f89c691892 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] No waiting events found dispatching network-vif-plugged-c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 891.170176] env[62368]: WARNING nova.compute.manager [req-18890692-ee3c-4029-91aa-5aefd4bf2c1a req-82edf882-b0db-40b7-80e8-31f89c691892 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received unexpected event network-vif-plugged-c06f647d-2b3f-4ee0-8221-d00fda253d7b for instance with vm_state building and task_state spawning. [ 891.183120] env[62368]: DEBUG nova.network.neutron [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Updating instance_info_cache with network_info: [{"id": "bd3ff77d-c1de-422a-81ca-03f3aa55bd9a", "address": "fa:16:3e:52:c8:2b", "network": {"id": "e89b0140-71be-4766-9ec4-c37da482a0e3", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2110723737-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb560320635f493fbe4308af1623e8ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd3ff77d-c1", "ovs_interfaceid": "bd3ff77d-c1de-422a-81ca-03f3aa55bd9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.223878] env[62368]: INFO nova.compute.manager [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Took 26.57 seconds to build instance. [ 891.247008] env[62368]: INFO nova.compute.manager [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Unrescuing [ 891.247326] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.247502] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.247675] env[62368]: DEBUG nova.network.neutron [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.343531] env[62368]: DEBUG nova.scheduler.client.report [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.375507] env[62368]: DEBUG nova.network.neutron [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Successfully updated port: c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 891.388226] env[62368]: INFO nova.compute.manager [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] instance snapshotting [ 891.391270] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28d30c0-cb82-4614-8e00-1e572c6f6bee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.412278] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dcb069-36c8-4643-86b3-6efd7026652a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.617200] env[62368]: DEBUG nova.network.neutron [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Port 11ce4c60-0df0-4c43-bbe4-6cc53717309b binding to destination host cpu-1 is already ACTIVE {{(pid=62368) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 891.686447] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "refresh_cache-207de202-ca50-4811-84c5-a0d4454d29f7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.686775] env[62368]: DEBUG nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Instance network_info: |[{"id": "bd3ff77d-c1de-422a-81ca-03f3aa55bd9a", "address": "fa:16:3e:52:c8:2b", "network": {"id": "e89b0140-71be-4766-9ec4-c37da482a0e3", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2110723737-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb560320635f493fbe4308af1623e8ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd3ff77d-c1", "ovs_interfaceid": "bd3ff77d-c1de-422a-81ca-03f3aa55bd9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 891.688063] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:c8:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f44b2fa3-6730-4b87-8839-947eff21213f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd3ff77d-c1de-422a-81ca-03f3aa55bd9a', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.697468] env[62368]: DEBUG oslo.service.loopingcall [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.698029] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 891.698324] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb503a40-c8b1-4e81-87e9-7f134f531e35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.721010] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.721010] env[62368]: value = "task-1198697" [ 891.721010] env[62368]: _type = "Task" [ 891.721010] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.725722] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5bbe5bf8-8f7f-4617-bac7-2824a9a790ee tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.088s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.733358] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198697, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.848607] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.851272] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.852s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.853020] env[62368]: INFO nova.compute.claims [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.877746] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.877892] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.878051] env[62368]: DEBUG nova.network.neutron [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.887068] env[62368]: INFO nova.scheduler.client.report [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Deleted allocations for instance a0cea538-b162-4504-ac34-803a2d5a8071 [ 891.924079] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 891.924420] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bbeefd55-b99c-43d5-9ca6-f48bc521f2a4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.941060] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 891.941060] env[62368]: value = "task-1198698" [ 891.941060] env[62368]: _type = "Task" [ 891.941060] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.948912] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198698, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.096667] env[62368]: DEBUG nova.network.neutron [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Updating instance_info_cache with network_info: [{"id": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "address": "fa:16:3e:fe:ae:1b", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b605458-f2", "ovs_interfaceid": "3b605458-f2bc-4c5e-8d9a-44fd80ab88bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.231674] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198697, 'name': CreateVM_Task, 'duration_secs': 0.367373} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.231844] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 892.232541] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.232727] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.233102] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.233371] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d71dba1-66cf-4576-ae3f-af8849aa61e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.239047] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 892.239047] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c35291-30c4-063e-3ede-2645ad6d4c82" [ 892.239047] env[62368]: _type = "Task" [ 892.239047] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.247477] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c35291-30c4-063e-3ede-2645ad6d4c82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.396775] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd484adb-f8e2-427b-ba97-e9f073704347 tempest-MigrationsAdminTest-1147189561 tempest-MigrationsAdminTest-1147189561-project-member] Lock "a0cea538-b162-4504-ac34-803a2d5a8071" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.828s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.419501] env[62368]: DEBUG nova.network.neutron [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 892.450481] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198698, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.601893] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "refresh_cache-b87094da-6258-469e-ab37-5557955ad3a7" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.603581] env[62368]: DEBUG nova.objects.instance [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lazy-loading 'flavor' on Instance uuid b87094da-6258-469e-ab37-5557955ad3a7 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.608585] env[62368]: DEBUG nova.network.neutron [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.641658] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.641658] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.641658] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.752121] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c35291-30c4-063e-3ede-2645ad6d4c82, 'name': SearchDatastore_Task, 'duration_secs': 0.012931} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.754023] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.754023] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.754023] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.754023] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.754459] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.755725] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d5a65bc-c688-42e0-97ab-bf27bf522420 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.768981] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.769168] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 892.769938] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4476dc94-f8cf-4c4b-8d11-2ad55dafbd9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.777042] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 892.777042] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]528f31a8-d13a-f158-6390-ef6b276edea1" [ 892.777042] env[62368]: _type = "Task" [ 892.777042] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.785666] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528f31a8-d13a-f158-6390-ef6b276edea1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.951580] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198698, 'name': CreateSnapshot_Task, 'duration_secs': 0.742444} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.955303] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 892.956600] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5699b05-f751-4de6-9fdb-2bc71192f690 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.117529] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33f25a1-a0c6-4929-965c-0d0744181b15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.122649] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.123019] env[62368]: DEBUG nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Instance network_info: |[{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 893.123674] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:9b:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c06f647d-2b3f-4ee0-8221-d00fda253d7b', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.134205] env[62368]: DEBUG oslo.service.loopingcall [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.135086] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 893.136274] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9e3c4e9-025c-4a3b-ae34-d233ddc91288 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.152883] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab59e93a-c42e-4d2c-88ff-37ad4da50776 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.177639] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 893.178660] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9322ecec-6927-4d86-92a5-98945f62564a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.185505] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91806218-cd7c-4f55-b0cb-e6b7523a3186 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.190438] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.190438] env[62368]: value = "task-1198699" [ 893.190438] env[62368]: _type = "Task" [ 893.190438] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.192636] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 893.192636] env[62368]: value = "task-1198700" [ 893.192636] env[62368]: _type = "Task" [ 893.192636] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.232344] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a69a9f-b663-4831-8471-1f7b233fd7be {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.236849] env[62368]: DEBUG nova.compute.manager [req-91fdb382-220d-4e5e-a046-622b42382d75 req-b5b1a8ca-716a-46bc-97d8-93982cba6d62 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.237413] env[62368]: DEBUG nova.compute.manager [req-91fdb382-220d-4e5e-a046-622b42382d75 req-b5b1a8ca-716a-46bc-97d8-93982cba6d62 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing instance network info cache due to event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 893.237413] env[62368]: DEBUG oslo_concurrency.lockutils [req-91fdb382-220d-4e5e-a046-622b42382d75 req-b5b1a8ca-716a-46bc-97d8-93982cba6d62 service nova] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.237590] env[62368]: DEBUG oslo_concurrency.lockutils [req-91fdb382-220d-4e5e-a046-622b42382d75 req-b5b1a8ca-716a-46bc-97d8-93982cba6d62 service nova] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.237795] env[62368]: DEBUG nova.network.neutron [req-91fdb382-220d-4e5e-a046-622b42382d75 req-b5b1a8ca-716a-46bc-97d8-93982cba6d62 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 893.245013] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198699, 'name': CreateVM_Task} progress is 15%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.250801] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198700, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.255476] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5720acb8-0f7a-40e2-97fc-34364509a3e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.273888] env[62368]: DEBUG nova.compute.provider_tree [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.289363] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528f31a8-d13a-f158-6390-ef6b276edea1, 'name': SearchDatastore_Task, 'duration_secs': 0.009206} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.290561] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe38f188-790d-4b28-8ffb-9620a855a1b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.297506] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 893.297506] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52331ade-2f5c-2058-4dba-dd54abe7ad8a" [ 893.297506] env[62368]: _type = "Task" [ 893.297506] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.311200] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52331ade-2f5c-2058-4dba-dd54abe7ad8a, 'name': SearchDatastore_Task, 'duration_secs': 0.011156} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.311503] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.311796] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 207de202-ca50-4811-84c5-a0d4454d29f7/207de202-ca50-4811-84c5-a0d4454d29f7.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 893.312115] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b67e971a-40ca-4b4c-9d67-aed15574ef02 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.320134] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 893.320134] env[62368]: value = "task-1198701" [ 893.320134] env[62368]: _type = "Task" [ 893.320134] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.329961] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.479871] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 893.480871] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cc18539a-87de-4636-a461-caaa281c8f61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.491469] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 893.491469] env[62368]: value = "task-1198702" [ 893.491469] env[62368]: _type = "Task" [ 893.491469] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.502604] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198702, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.714033] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198699, 'name': CreateVM_Task, 'duration_secs': 0.422867} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.714508] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198700, 'name': PowerOffVM_Task, 'duration_secs': 0.261338} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.714805] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 893.715223] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 893.721377] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 893.722936] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.723595] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.723595] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 893.723860] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89daef9f-05e7-49d3-a2d2-277a06a9a633 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.738569] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.738754] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.738936] env[62368]: DEBUG nova.network.neutron [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.740299] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-597c7c4e-5da7-4b77-98dd-37844b3fcb17 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.751621] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 893.751621] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52587fb1-3b13-8cc8-7d93-3db341060f08" [ 893.751621] env[62368]: _type = "Task" [ 893.751621] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.753754] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 893.753754] env[62368]: value = "task-1198703" [ 893.753754] env[62368]: _type = "Task" [ 893.753754] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.770082] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52587fb1-3b13-8cc8-7d93-3db341060f08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.774840] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198703, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.777127] env[62368]: DEBUG nova.scheduler.client.report [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.832928] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198701, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.962968] env[62368]: DEBUG nova.compute.manager [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Stashing vm_state: active {{(pid=62368) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 894.006965] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198702, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.130118] env[62368]: DEBUG nova.network.neutron [req-91fdb382-220d-4e5e-a046-622b42382d75 req-b5b1a8ca-716a-46bc-97d8-93982cba6d62 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updated VIF entry in instance network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 894.130520] env[62368]: DEBUG nova.network.neutron [req-91fdb382-220d-4e5e-a046-622b42382d75 req-b5b1a8ca-716a-46bc-97d8-93982cba6d62 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.269283] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52587fb1-3b13-8cc8-7d93-3db341060f08, 'name': SearchDatastore_Task, 'duration_secs': 0.06089} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.272938] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.273227] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.273475] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.273620] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.273799] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.274104] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198703, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.274359] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-217569da-3b2d-47e2-99f9-937846281870 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.282406] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.282958] env[62368]: DEBUG nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 894.287644] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.966s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.289488] env[62368]: INFO nova.compute.claims [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.292136] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.292326] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 894.296022] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29e4d762-88a6-4c03-b168-75cee9e86944 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.302348] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 894.302348] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52680faa-e65c-5f17-ea66-d0f689376027" [ 894.302348] env[62368]: _type = "Task" [ 894.302348] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.311887] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52680faa-e65c-5f17-ea66-d0f689376027, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.332307] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539729} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.333504] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 207de202-ca50-4811-84c5-a0d4454d29f7/207de202-ca50-4811-84c5-a0d4454d29f7.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 894.333742] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.334022] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-585a9dcf-8a64-4bf8-87a7-91994ccd01ef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.342858] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 894.342858] env[62368]: value = "task-1198704" [ 894.342858] env[62368]: _type = "Task" [ 894.342858] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.351908] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.487920] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.514343] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198702, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.569138] env[62368]: DEBUG nova.network.neutron [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.634742] env[62368]: DEBUG oslo_concurrency.lockutils [req-91fdb382-220d-4e5e-a046-622b42382d75 req-b5b1a8ca-716a-46bc-97d8-93982cba6d62 service nova] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.771933] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198703, 'name': ReconfigVM_Task, 'duration_secs': 0.543548} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.772314] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 894.772515] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 894.772817] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41ceffd9-0e5d-4e1e-9620-880ef1eaad7c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.780210] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 894.780210] env[62368]: value = "task-1198705" [ 894.780210] env[62368]: _type = "Task" [ 894.780210] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.788933] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198705, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.795111] env[62368]: DEBUG nova.compute.utils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.799459] env[62368]: DEBUG nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 894.799628] env[62368]: DEBUG nova.network.neutron [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 894.813349] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52680faa-e65c-5f17-ea66-d0f689376027, 'name': SearchDatastore_Task, 'duration_secs': 0.010284} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.815590] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2f5920e-9486-4221-9a0b-31b6813a0b64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.820742] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 894.820742] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]523b0061-3c56-f989-32de-ac73748467b3" [ 894.820742] env[62368]: _type = "Task" [ 894.820742] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.831413] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523b0061-3c56-f989-32de-ac73748467b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.856128] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074571} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.856128] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.856128] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1189fdd2-56d3-42cc-9a2a-ce469583a7d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.878989] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 207de202-ca50-4811-84c5-a0d4454d29f7/207de202-ca50-4811-84c5-a0d4454d29f7.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.880825] env[62368]: DEBUG nova.policy [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87c366f937e2485c8b0dc317d6acb8f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '430e72f3b08a4207a2fb4465ff160714', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.882487] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2213d4d-9dbf-4681-aa7a-00f1c44a997c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.906642] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 894.906642] env[62368]: value = "task-1198706" [ 894.906642] env[62368]: _type = "Task" [ 894.906642] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.915314] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198706, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.012907] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198702, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.071845] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.295072] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198705, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.300068] env[62368]: DEBUG nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 895.335717] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523b0061-3c56-f989-32de-ac73748467b3, 'name': SearchDatastore_Task, 'duration_secs': 0.009651} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.337064] env[62368]: DEBUG nova.network.neutron [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Successfully created port: 41e7d4b7-38bf-4270-812a-ac72f69e40d6 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 895.342733] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.343416] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] a5cbadbd-20dd-4514-8867-20243af5db0c/a5cbadbd-20dd-4514-8867-20243af5db0c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 895.344231] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bec864aa-1eea-4857-a886-b2510ac0368f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.354730] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 895.354730] env[62368]: value = "task-1198707" [ 895.354730] env[62368]: _type = "Task" [ 895.354730] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.372766] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198707, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.423554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.423831] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.425686] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198706, 'name': ReconfigVM_Task, 'duration_secs': 0.306757} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.427738] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 207de202-ca50-4811-84c5-a0d4454d29f7/207de202-ca50-4811-84c5-a0d4454d29f7.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.427738] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35a7608f-c8c1-4264-8616-a01db90e44eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.434602] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 895.434602] env[62368]: value = "task-1198708" [ 895.434602] env[62368]: _type = "Task" [ 895.434602] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.449214] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198708, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.512114] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198702, 'name': CloneVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.602769] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c8e96b-945c-486c-8bcc-5e2e9236507d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.635232] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd5fe53-6083-4f94-956b-40106a88e819 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.645074] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance '13765305-2e55-4ee8-9a6f-4ae5ee724367' progress to 83 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 895.655665] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd654c2-8bf7-4422-8490-e91d91ec506b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.664780] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9497d80-2f96-4d41-b237-39e7309d1899 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.700971] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1865d238-b72b-427d-a52e-ff703f312ef0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.710968] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efc99e9-9751-4a4c-b4cc-60c225637f0c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.728710] env[62368]: DEBUG nova.compute.provider_tree [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.793020] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198705, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.866475] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198707, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.926496] env[62368]: DEBUG nova.compute.manager [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 895.945998] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198708, 'name': Rename_Task, 'duration_secs': 0.35717} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.946237] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.946541] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39426c81-2932-403d-847c-b95a939f241f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.954680] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 895.954680] env[62368]: value = "task-1198709" [ 895.954680] env[62368]: _type = "Task" [ 895.954680] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.966531] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198709, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.008201] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198702, 'name': CloneVM_Task, 'duration_secs': 2.122485} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.008563] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Created linked-clone VM from snapshot [ 896.009337] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d7087f-34c8-4dd2-8391-9be5926a59d4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.017388] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Uploading image 77367ccf-9c4b-409b-8db5-b72274225af7 {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 896.029481] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 896.030369] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-81beb466-6a8e-48c1-a411-5775e8196506 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.039345] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 896.039345] env[62368]: value = "task-1198710" [ 896.039345] env[62368]: _type = "Task" [ 896.039345] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.049086] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198710, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.153064] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdd4112-5c96-4336-b839-cd485bcac549 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance '13765305-2e55-4ee8-9a6f-4ae5ee724367' progress to 100 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 896.234721] env[62368]: DEBUG nova.scheduler.client.report [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.291963] env[62368]: DEBUG oslo_vmware.api [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198705, 'name': PowerOnVM_Task, 'duration_secs': 1.307493} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.292278] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.292524] env[62368]: DEBUG nova.compute.manager [None req-c0e8697c-c5ba-480d-85fa-be39cf0ae4c5 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.293313] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02356a2e-ca21-4d9c-af13-9c3c0db78782 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.312997] env[62368]: DEBUG nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 896.338702] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.339084] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.339191] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.339384] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.339540] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.339713] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.339954] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.340140] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.340332] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.340503] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.340680] env[62368]: DEBUG nova.virt.hardware [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.341833] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd02a8c2-3b0b-456e-8642-c9cea3c05d99 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.350550] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4ca59b-be40-425a-9802-45987d7bec2a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.372217] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198707, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520962} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.372521] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] a5cbadbd-20dd-4514-8867-20243af5db0c/a5cbadbd-20dd-4514-8867-20243af5db0c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 896.372711] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.372956] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3464ba9-2b9c-4bd2-ab02-7d0a18531168 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.380246] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 896.380246] env[62368]: value = "task-1198711" [ 896.380246] env[62368]: _type = "Task" [ 896.380246] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.389309] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198711, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.450681] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.466927] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198709, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.552587] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198710, 'name': Destroy_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.739204] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.739876] env[62368]: DEBUG nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.742579] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.601s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.742757] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.742898] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 896.743197] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.337s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.746667] env[62368]: INFO nova.compute.claims [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.750207] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eeabcff-b2a5-4d82-848b-766540fd5a15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.760015] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8667867-7fb7-4b62-9ba7-5e47c6e952b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.775507] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f235e47-26c5-4d2c-bef6-6e1419290ca4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.783461] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9771bd6-ef69-46a6-9bbe-1f55dac69141 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.818655] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179624MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 896.818808] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.894630] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198711, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086853} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.895040] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.895881] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74c07d0-a196-4e3c-814d-6d68da9337a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.922956] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] a5cbadbd-20dd-4514-8867-20243af5db0c/a5cbadbd-20dd-4514-8867-20243af5db0c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.923788] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9badbc4d-6d25-4ac6-8623-953c1f9dff9c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.946477] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 896.946477] env[62368]: value = "task-1198712" [ 896.946477] env[62368]: _type = "Task" [ 896.946477] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.957163] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198712, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.967090] env[62368]: DEBUG oslo_vmware.api [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198709, 'name': PowerOnVM_Task, 'duration_secs': 0.71756} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.967427] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.967646] env[62368]: INFO nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Took 9.13 seconds to spawn the instance on the hypervisor. [ 896.967840] env[62368]: DEBUG nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.968643] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127163b4-bb12-42bb-8937-fa464eda2415 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.053427] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198710, 'name': Destroy_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.183622] env[62368]: DEBUG nova.compute.manager [req-68e7e32f-df31-4b35-b38b-83e3b08ec229 req-f3cfd2c0-fa05-425e-b279-113ac2d77bc6 service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Received event network-vif-plugged-41e7d4b7-38bf-4270-812a-ac72f69e40d6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.186027] env[62368]: DEBUG oslo_concurrency.lockutils [req-68e7e32f-df31-4b35-b38b-83e3b08ec229 req-f3cfd2c0-fa05-425e-b279-113ac2d77bc6 service nova] Acquiring lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.186027] env[62368]: DEBUG oslo_concurrency.lockutils [req-68e7e32f-df31-4b35-b38b-83e3b08ec229 req-f3cfd2c0-fa05-425e-b279-113ac2d77bc6 service nova] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.186027] env[62368]: DEBUG oslo_concurrency.lockutils [req-68e7e32f-df31-4b35-b38b-83e3b08ec229 req-f3cfd2c0-fa05-425e-b279-113ac2d77bc6 service nova] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.186027] env[62368]: DEBUG nova.compute.manager [req-68e7e32f-df31-4b35-b38b-83e3b08ec229 req-f3cfd2c0-fa05-425e-b279-113ac2d77bc6 service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] No waiting events found dispatching network-vif-plugged-41e7d4b7-38bf-4270-812a-ac72f69e40d6 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 897.186027] env[62368]: WARNING nova.compute.manager [req-68e7e32f-df31-4b35-b38b-83e3b08ec229 req-f3cfd2c0-fa05-425e-b279-113ac2d77bc6 service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Received unexpected event network-vif-plugged-41e7d4b7-38bf-4270-812a-ac72f69e40d6 for instance with vm_state building and task_state spawning. [ 897.195455] env[62368]: DEBUG nova.network.neutron [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Successfully updated port: 41e7d4b7-38bf-4270-812a-ac72f69e40d6 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.252675] env[62368]: DEBUG nova.compute.utils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.256515] env[62368]: DEBUG nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.256704] env[62368]: DEBUG nova.network.neutron [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 897.321932] env[62368]: DEBUG nova.policy [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c66c425264424e3eaa37c8058e62cb30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '026a3176abc34001ab3506445226777f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 897.462108] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198712, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.489825] env[62368]: INFO nova.compute.manager [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Took 31.61 seconds to build instance. [ 897.552676] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198710, 'name': Destroy_Task, 'duration_secs': 1.081455} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.553009] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Destroyed the VM [ 897.553272] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 897.553535] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-130143bd-9b38-4c28-8ae5-69966b138ebe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.562387] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 897.562387] env[62368]: value = "task-1198713" [ 897.562387] env[62368]: _type = "Task" [ 897.562387] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.575225] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198713, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.697942] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "refresh_cache-14c358a0-434d-472a-a382-b5b0e9bf2bd8" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.698138] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "refresh_cache-14c358a0-434d-472a-a382-b5b0e9bf2bd8" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.698251] env[62368]: DEBUG nova.network.neutron [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.760950] env[62368]: DEBUG nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.816592] env[62368]: DEBUG nova.network.neutron [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Successfully created port: b8138e3d-417d-4a57-8c35-ac1d61118665 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.958711] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "207de202-ca50-4811-84c5-a0d4454d29f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.965671] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198712, 'name': ReconfigVM_Task, 'duration_secs': 0.70399} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.965972] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Reconfigured VM instance instance-00000053 to attach disk [datastore2] a5cbadbd-20dd-4514-8867-20243af5db0c/a5cbadbd-20dd-4514-8867-20243af5db0c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.966656] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fb4d2ff-b343-4e55-a65d-9c600522a069 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.975229] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 897.975229] env[62368]: value = "task-1198714" [ 897.975229] env[62368]: _type = "Task" [ 897.975229] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.990115] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198714, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.996970] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b02b6b55-5666-4024-9376-1f44933b993f tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "207de202-ca50-4811-84c5-a0d4454d29f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.124s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.997503] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "207de202-ca50-4811-84c5-a0d4454d29f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.039s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.997750] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "207de202-ca50-4811-84c5-a0d4454d29f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.999937] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "207de202-ca50-4811-84c5-a0d4454d29f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.999937] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "207de202-ca50-4811-84c5-a0d4454d29f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.001347] env[62368]: INFO nova.compute.manager [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Terminating instance [ 898.003782] env[62368]: DEBUG nova.compute.manager [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 898.003782] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 898.004703] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18d3229-6106-4de5-ba59-8fe33b5adf7c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.014320] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 898.017444] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f6d7350-60d0-417a-9ad3-1f946a112390 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.027404] env[62368]: DEBUG oslo_vmware.api [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 898.027404] env[62368]: value = "task-1198715" [ 898.027404] env[62368]: _type = "Task" [ 898.027404] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.040812] env[62368]: DEBUG oslo_vmware.api [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198715, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.073462] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70568db2-ff31-4ac1-9793-e2c84e149e64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.083548] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198713, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.087771] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d93a4e-82e5-45da-950b-cd1ea4af5d49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.133969] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3649f0-20eb-4b02-82c6-4f0f3ac5a77f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.148549] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5719bfb8-4216-4980-923a-f0e18676843d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.169861] env[62368]: DEBUG nova.compute.provider_tree [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.303687] env[62368]: DEBUG nova.network.neutron [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 898.491270] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198714, 'name': Rename_Task, 'duration_secs': 0.307431} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.493691] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 898.494060] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ae2653c-f207-465b-adc9-7b95286be781 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.513022] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 898.513022] env[62368]: value = "task-1198716" [ 898.513022] env[62368]: _type = "Task" [ 898.513022] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.521994] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.545860] env[62368]: DEBUG oslo_vmware.api [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198715, 'name': PowerOffVM_Task, 'duration_secs': 0.255794} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.546815] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 898.550194] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 898.550194] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-168fceba-795d-4e58-bd21-638495e63b24 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.580322] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198713, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.616598] env[62368]: DEBUG nova.network.neutron [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Updating instance_info_cache with network_info: [{"id": "41e7d4b7-38bf-4270-812a-ac72f69e40d6", "address": "fa:16:3e:fc:57:3a", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41e7d4b7-38", "ovs_interfaceid": "41e7d4b7-38bf-4270-812a-ac72f69e40d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.636120] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 898.636443] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 898.637017] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleting the datastore file [datastore2] 207de202-ca50-4811-84c5-a0d4454d29f7 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.637500] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8d7a779-6132-4fdc-bf43-54dbb0ef76c9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.647973] env[62368]: DEBUG oslo_vmware.api [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for the task: (returnval){ [ 898.647973] env[62368]: value = "task-1198718" [ 898.647973] env[62368]: _type = "Task" [ 898.647973] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.660341] env[62368]: DEBUG oslo_vmware.api [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198718, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.672691] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.673039] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.673250] env[62368]: DEBUG nova.compute.manager [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Going to confirm migration 2 {{(pid=62368) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 898.677328] env[62368]: DEBUG nova.scheduler.client.report [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.775199] env[62368]: DEBUG nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.809234] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.809501] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.809663] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.809852] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.810013] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.810761] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.811007] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.811194] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.811374] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.811547] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.811729] env[62368]: DEBUG nova.virt.hardware [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.812657] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad3df84-c950-4702-8e2c-8b208f808aad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.822984] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927f615f-16bc-4af9-8680-25adbb6dc98d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.026213] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198716, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.076915] env[62368]: DEBUG oslo_vmware.api [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198713, 'name': RemoveSnapshot_Task, 'duration_secs': 1.072719} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.077272] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 899.120625] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "refresh_cache-14c358a0-434d-472a-a382-b5b0e9bf2bd8" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.120718] env[62368]: DEBUG nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Instance network_info: |[{"id": "41e7d4b7-38bf-4270-812a-ac72f69e40d6", "address": "fa:16:3e:fc:57:3a", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41e7d4b7-38", "ovs_interfaceid": "41e7d4b7-38bf-4270-812a-ac72f69e40d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.121373] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:57:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41e7d4b7-38bf-4270-812a-ac72f69e40d6', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.139435] env[62368]: DEBUG oslo.service.loopingcall [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.139853] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 899.140335] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d66d0d8-9c07-43e3-97bc-65c766e7d213 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.187748] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.187748] env[62368]: DEBUG nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.197768] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.710s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.200841] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.200841] env[62368]: value = "task-1198719" [ 899.200841] env[62368]: _type = "Task" [ 899.200841] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.216084] env[62368]: DEBUG oslo_vmware.api [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Task: {'id': task-1198718, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.525351} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.218417] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.218796] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 899.219117] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 899.219503] env[62368]: INFO nova.compute.manager [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Took 1.22 seconds to destroy the instance on the hypervisor. [ 899.219784] env[62368]: DEBUG oslo.service.loopingcall [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.220458] env[62368]: DEBUG nova.compute.manager [-] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 899.220608] env[62368]: DEBUG nova.network.neutron [-] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 899.228686] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198719, 'name': CreateVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.233512] env[62368]: DEBUG nova.compute.manager [req-8fdda276-9d86-426b-b2df-88ccf41c4a46 req-b120f37d-0675-4035-9dfb-0dd194b50c34 service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Received event network-changed-41e7d4b7-38bf-4270-812a-ac72f69e40d6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.233512] env[62368]: DEBUG nova.compute.manager [req-8fdda276-9d86-426b-b2df-88ccf41c4a46 req-b120f37d-0675-4035-9dfb-0dd194b50c34 service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Refreshing instance network info cache due to event network-changed-41e7d4b7-38bf-4270-812a-ac72f69e40d6. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 899.233512] env[62368]: DEBUG oslo_concurrency.lockutils [req-8fdda276-9d86-426b-b2df-88ccf41c4a46 req-b120f37d-0675-4035-9dfb-0dd194b50c34 service nova] Acquiring lock "refresh_cache-14c358a0-434d-472a-a382-b5b0e9bf2bd8" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.233512] env[62368]: DEBUG oslo_concurrency.lockutils [req-8fdda276-9d86-426b-b2df-88ccf41c4a46 req-b120f37d-0675-4035-9dfb-0dd194b50c34 service nova] Acquired lock "refresh_cache-14c358a0-434d-472a-a382-b5b0e9bf2bd8" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.233833] env[62368]: DEBUG nova.network.neutron [req-8fdda276-9d86-426b-b2df-88ccf41c4a46 req-b120f37d-0675-4035-9dfb-0dd194b50c34 service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Refreshing network info cache for port 41e7d4b7-38bf-4270-812a-ac72f69e40d6 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 899.258469] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.258700] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.258949] env[62368]: DEBUG nova.network.neutron [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 899.259274] env[62368]: DEBUG nova.objects.instance [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'info_cache' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.534790] env[62368]: DEBUG oslo_vmware.api [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198716, 'name': PowerOnVM_Task, 'duration_secs': 0.855485} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.535257] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 899.536423] env[62368]: INFO nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Took 9.38 seconds to spawn the instance on the hypervisor. [ 899.536656] env[62368]: DEBUG nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.537869] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72351f93-e8c8-4b5d-a86a-a261a5b1897f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.583539] env[62368]: WARNING nova.compute.manager [None req-d93ce365-c816-4a43-a3b0-e025a2ec2138 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Image not found during snapshot: nova.exception.ImageNotFound: Image 77367ccf-9c4b-409b-8db5-b72274225af7 could not be found. [ 899.703199] env[62368]: DEBUG nova.compute.utils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.708151] env[62368]: INFO nova.compute.claims [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.712621] env[62368]: DEBUG nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 899.712621] env[62368]: DEBUG nova.network.neutron [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 899.730815] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198719, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.746056] env[62368]: DEBUG nova.compute.manager [req-e88ad357-33ff-4281-ab96-e472b5fe25b2 req-846018c0-392c-4821-a5f2-faf324c9aa2f service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Received event network-vif-deleted-bd3ff77d-c1de-422a-81ca-03f3aa55bd9a {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.746386] env[62368]: INFO nova.compute.manager [req-e88ad357-33ff-4281-ab96-e472b5fe25b2 req-846018c0-392c-4821-a5f2-faf324c9aa2f service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Neutron deleted interface bd3ff77d-c1de-422a-81ca-03f3aa55bd9a; detaching it from the instance and deleting it from the info cache [ 899.746519] env[62368]: DEBUG nova.network.neutron [req-e88ad357-33ff-4281-ab96-e472b5fe25b2 req-846018c0-392c-4821-a5f2-faf324c9aa2f service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.812929] env[62368]: DEBUG nova.policy [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7dceb8c8f8b4c4dbdde4d097009d429', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b70fcc6664f47d7b55447210851c4cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 899.877817] env[62368]: DEBUG nova.compute.manager [req-1c41f8af-9de9-4509-8c56-457e6396f40c req-919f1fec-77a8-4bd9-a524-8c5116a86c92 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Received event network-vif-plugged-b8138e3d-417d-4a57-8c35-ac1d61118665 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.878155] env[62368]: DEBUG oslo_concurrency.lockutils [req-1c41f8af-9de9-4509-8c56-457e6396f40c req-919f1fec-77a8-4bd9-a524-8c5116a86c92 service nova] Acquiring lock "ad21bf3a-068e-44a7-b58f-2978a631b972-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.878314] env[62368]: DEBUG oslo_concurrency.lockutils [req-1c41f8af-9de9-4509-8c56-457e6396f40c req-919f1fec-77a8-4bd9-a524-8c5116a86c92 service nova] Lock "ad21bf3a-068e-44a7-b58f-2978a631b972-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.878489] env[62368]: DEBUG oslo_concurrency.lockutils [req-1c41f8af-9de9-4509-8c56-457e6396f40c req-919f1fec-77a8-4bd9-a524-8c5116a86c92 service nova] Lock "ad21bf3a-068e-44a7-b58f-2978a631b972-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.878658] env[62368]: DEBUG nova.compute.manager [req-1c41f8af-9de9-4509-8c56-457e6396f40c req-919f1fec-77a8-4bd9-a524-8c5116a86c92 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] No waiting events found dispatching network-vif-plugged-b8138e3d-417d-4a57-8c35-ac1d61118665 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.878828] env[62368]: WARNING nova.compute.manager [req-1c41f8af-9de9-4509-8c56-457e6396f40c req-919f1fec-77a8-4bd9-a524-8c5116a86c92 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Received unexpected event network-vif-plugged-b8138e3d-417d-4a57-8c35-ac1d61118665 for instance with vm_state building and task_state spawning. [ 900.072774] env[62368]: INFO nova.compute.manager [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Took 28.76 seconds to build instance. [ 900.149460] env[62368]: DEBUG nova.network.neutron [-] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.159036] env[62368]: DEBUG nova.network.neutron [req-8fdda276-9d86-426b-b2df-88ccf41c4a46 req-b120f37d-0675-4035-9dfb-0dd194b50c34 service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Updated VIF entry in instance network info cache for port 41e7d4b7-38bf-4270-812a-ac72f69e40d6. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 900.159036] env[62368]: DEBUG nova.network.neutron [req-8fdda276-9d86-426b-b2df-88ccf41c4a46 req-b120f37d-0675-4035-9dfb-0dd194b50c34 service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Updating instance_info_cache with network_info: [{"id": "41e7d4b7-38bf-4270-812a-ac72f69e40d6", "address": "fa:16:3e:fc:57:3a", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41e7d4b7-38", "ovs_interfaceid": "41e7d4b7-38bf-4270-812a-ac72f69e40d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.224689] env[62368]: INFO nova.compute.resource_tracker [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating resource usage from migration db16e018-9b81-40e4-8b6b-0ac690c9f100 [ 900.227826] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198719, 'name': CreateVM_Task, 'duration_secs': 0.760945} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.229545] env[62368]: DEBUG nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.235067] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 900.235067] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.235067] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.235067] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.235067] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-194a97ce-3ff0-4e10-9b45-a7623157f092 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.238824] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 900.238824] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ad8f0d-73cd-6962-716e-c5358c5167d9" [ 900.238824] env[62368]: _type = "Task" [ 900.238824] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.251347] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ad8f0d-73cd-6962-716e-c5358c5167d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.251948] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8d6070d-c0e1-4c13-9751-cd4c76351555 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.265516] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7ed36d-46f1-430a-997f-103ada180b1b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.305726] env[62368]: DEBUG nova.compute.manager [req-e88ad357-33ff-4281-ab96-e472b5fe25b2 req-846018c0-392c-4821-a5f2-faf324c9aa2f service nova] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Detach interface failed, port_id=bd3ff77d-c1de-422a-81ca-03f3aa55bd9a, reason: Instance 207de202-ca50-4811-84c5-a0d4454d29f7 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 900.464037] env[62368]: DEBUG nova.network.neutron [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Successfully updated port: b8138e3d-417d-4a57-8c35-ac1d61118665 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.523225] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219af148-04a2-4f48-958f-57a555f3b0d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.526760] env[62368]: DEBUG nova.network.neutron [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Successfully created port: 58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.537719] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5e57bc-934e-44d8-891f-477e0933841a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.571863] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a67eb97-b39d-4a90-b482-c95f2f801a09 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.580064] env[62368]: DEBUG oslo_concurrency.lockutils [None req-dc418345-301a-4cf6-8632-f8643b7c289b tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.276s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.581404] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529e3082-4b70-4a6c-99f8-9ac421e0b3fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.596589] env[62368]: DEBUG nova.compute.provider_tree [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.653265] env[62368]: INFO nova.compute.manager [-] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Took 1.43 seconds to deallocate network for instance. [ 900.662616] env[62368]: DEBUG oslo_concurrency.lockutils [req-8fdda276-9d86-426b-b2df-88ccf41c4a46 req-b120f37d-0675-4035-9dfb-0dd194b50c34 service nova] Releasing lock "refresh_cache-14c358a0-434d-472a-a382-b5b0e9bf2bd8" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.751396] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ad8f0d-73cd-6962-716e-c5358c5167d9, 'name': SearchDatastore_Task, 'duration_secs': 0.015628} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.751464] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.751976] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.751976] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.752117] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.752262] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.752498] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b1cfda3-18ed-406a-8adf-0e5dc3609469 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.762371] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.762633] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 900.764766] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9558dfa-704f-4f98-b8ff-8eb570a93e88 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.769538] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 900.769538] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]523ebd64-6746-3f8f-0df5-28e1db96f1c0" [ 900.769538] env[62368]: _type = "Task" [ 900.769538] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.777510] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523ebd64-6746-3f8f-0df5-28e1db96f1c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.914678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.914946] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.973612] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "refresh_cache-ad21bf3a-068e-44a7-b58f-2978a631b972" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.973612] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "refresh_cache-ad21bf3a-068e-44a7-b58f-2978a631b972" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.973612] env[62368]: DEBUG nova.network.neutron [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.052983] env[62368]: DEBUG nova.network.neutron [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.100148] env[62368]: DEBUG nova.scheduler.client.report [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.163218] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.253325] env[62368]: DEBUG nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.288172] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]523ebd64-6746-3f8f-0df5-28e1db96f1c0, 'name': SearchDatastore_Task, 'duration_secs': 0.01031} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.290615] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.291108] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.291411] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.291979] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.292477] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.292949] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.293288] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.293767] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.294331] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.294737] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.295059] env[62368]: DEBUG nova.virt.hardware [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.302195] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c572c41-b9c7-408d-a52d-a592de61ead9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.302551] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10dfce30-d783-46d2-93a3-a5cbcffde0cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.311422] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 901.311422] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5293900a-54f2-775b-d7be-6fcebb7f51e5" [ 901.311422] env[62368]: _type = "Task" [ 901.311422] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.319050] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e93246e-375e-43f7-ab45-f08d55d2ee27 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.329158] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5293900a-54f2-775b-d7be-6fcebb7f51e5, 'name': SearchDatastore_Task, 'duration_secs': 0.011196} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.337111] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.337404] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8/14c358a0-434d-472a-a382-b5b0e9bf2bd8.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 901.337961] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43880802-cc96-4411-91d8-3443238f19f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.344264] env[62368]: DEBUG nova.compute.manager [req-c5fc320e-96e0-48fe-bee7-51a9715e1014 req-19f4876e-20f5-41e7-84e0-bb0eb7e6bb93 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Received event network-changed-b8138e3d-417d-4a57-8c35-ac1d61118665 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.344425] env[62368]: DEBUG nova.compute.manager [req-c5fc320e-96e0-48fe-bee7-51a9715e1014 req-19f4876e-20f5-41e7-84e0-bb0eb7e6bb93 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Refreshing instance network info cache due to event network-changed-b8138e3d-417d-4a57-8c35-ac1d61118665. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.344477] env[62368]: DEBUG oslo_concurrency.lockutils [req-c5fc320e-96e0-48fe-bee7-51a9715e1014 req-19f4876e-20f5-41e7-84e0-bb0eb7e6bb93 service nova] Acquiring lock "refresh_cache-ad21bf3a-068e-44a7-b58f-2978a631b972" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.346858] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 901.346858] env[62368]: value = "task-1198720" [ 901.346858] env[62368]: _type = "Task" [ 901.346858] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.358674] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198720, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.419265] env[62368]: DEBUG nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 901.523028] env[62368]: DEBUG nova.network.neutron [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 901.556594] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.557222] env[62368]: DEBUG nova.objects.instance [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'migration_context' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.610161] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.412s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.612620] env[62368]: INFO nova.compute.manager [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Migrating [ 901.620834] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.170s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.623616] env[62368]: INFO nova.compute.claims [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.662144] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "55eecf3d-501c-490d-bddd-0211fd082841" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.662581] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "55eecf3d-501c-490d-bddd-0211fd082841" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.664805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "55eecf3d-501c-490d-bddd-0211fd082841-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.665213] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "55eecf3d-501c-490d-bddd-0211fd082841-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.665436] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "55eecf3d-501c-490d-bddd-0211fd082841-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.668832] env[62368]: INFO nova.compute.manager [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Terminating instance [ 901.672269] env[62368]: DEBUG nova.compute.manager [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.672492] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 901.675524] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada40295-3ba5-4a38-95f9-3692303b847b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.687081] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 901.687435] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-636fb504-18cb-4434-b9e2-32119cffc456 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.697176] env[62368]: DEBUG oslo_vmware.api [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 901.697176] env[62368]: value = "task-1198721" [ 901.697176] env[62368]: _type = "Task" [ 901.697176] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.707476] env[62368]: DEBUG oslo_vmware.api [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198721, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.774805] env[62368]: DEBUG nova.network.neutron [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Updating instance_info_cache with network_info: [{"id": "b8138e3d-417d-4a57-8c35-ac1d61118665", "address": "fa:16:3e:48:d3:0d", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8138e3d-41", "ovs_interfaceid": "b8138e3d-417d-4a57-8c35-ac1d61118665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.858543] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198720, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.950726] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.062457] env[62368]: DEBUG nova.objects.base [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Object Instance<13765305-2e55-4ee8-9a6f-4ae5ee724367> lazy-loaded attributes: info_cache,migration_context {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 902.062527] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2674eb-b385-4585-8716-af9f2d5162ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.099172] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e930e9dc-73a9-4dce-a2c2-84b29f894028 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.106157] env[62368]: DEBUG oslo_vmware.api [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 902.106157] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]528d6d9f-a374-38df-bbbe-1730b685831e" [ 902.106157] env[62368]: _type = "Task" [ 902.106157] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.125086] env[62368]: DEBUG oslo_vmware.api [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528d6d9f-a374-38df-bbbe-1730b685831e, 'name': SearchDatastore_Task, 'duration_secs': 0.008156} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.125496] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.144931] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.145171] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.145357] env[62368]: DEBUG nova.network.neutron [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 902.216851] env[62368]: DEBUG oslo_vmware.api [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198721, 'name': PowerOffVM_Task, 'duration_secs': 0.316508} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.217151] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 902.217340] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 902.217633] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d2cab01-0a6c-4edf-b724-060cb5f10a0e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.280659] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "refresh_cache-ad21bf3a-068e-44a7-b58f-2978a631b972" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.280659] env[62368]: DEBUG nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Instance network_info: |[{"id": "b8138e3d-417d-4a57-8c35-ac1d61118665", "address": "fa:16:3e:48:d3:0d", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8138e3d-41", "ovs_interfaceid": "b8138e3d-417d-4a57-8c35-ac1d61118665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 902.280659] env[62368]: DEBUG oslo_concurrency.lockutils [req-c5fc320e-96e0-48fe-bee7-51a9715e1014 req-19f4876e-20f5-41e7-84e0-bb0eb7e6bb93 service nova] Acquired lock "refresh_cache-ad21bf3a-068e-44a7-b58f-2978a631b972" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.280659] env[62368]: DEBUG nova.network.neutron [req-c5fc320e-96e0-48fe-bee7-51a9715e1014 req-19f4876e-20f5-41e7-84e0-bb0eb7e6bb93 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Refreshing network info cache for port b8138e3d-417d-4a57-8c35-ac1d61118665 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.282830] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:d3:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8f40f5c4-c146-449c-884d-6f884dcf2acf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8138e3d-417d-4a57-8c35-ac1d61118665', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.296416] env[62368]: DEBUG oslo.service.loopingcall [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.297513] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 902.297760] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7d1f0cc-7584-400f-b8f5-2d7110a23776 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.315467] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 902.315710] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 902.315886] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleting the datastore file [datastore2] 55eecf3d-501c-490d-bddd-0211fd082841 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.316621] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-164776eb-e632-4b63-af92-f09008b9431b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.324864] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.324864] env[62368]: value = "task-1198723" [ 902.324864] env[62368]: _type = "Task" [ 902.324864] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.329636] env[62368]: DEBUG oslo_vmware.api [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for the task: (returnval){ [ 902.329636] env[62368]: value = "task-1198724" [ 902.329636] env[62368]: _type = "Task" [ 902.329636] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.339343] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198723, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.342336] env[62368]: DEBUG oslo_vmware.api [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198724, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.358374] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198720, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518667} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.358693] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8/14c358a0-434d-472a-a382-b5b0e9bf2bd8.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 902.358947] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 902.359244] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e13ba320-382e-46f1-bb47-33bb0443db73 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.369478] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 902.369478] env[62368]: value = "task-1198725" [ 902.369478] env[62368]: _type = "Task" [ 902.369478] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.380432] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198725, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.632510] env[62368]: DEBUG nova.compute.manager [req-38450e22-ec4f-4c40-ac78-037346a54bb5 req-5178ecef-0a77-47df-a8b6-bfd9d2557ff6 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Received event network-vif-plugged-58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.632510] env[62368]: DEBUG oslo_concurrency.lockutils [req-38450e22-ec4f-4c40-ac78-037346a54bb5 req-5178ecef-0a77-47df-a8b6-bfd9d2557ff6 service nova] Acquiring lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.632510] env[62368]: DEBUG oslo_concurrency.lockutils [req-38450e22-ec4f-4c40-ac78-037346a54bb5 req-5178ecef-0a77-47df-a8b6-bfd9d2557ff6 service nova] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.632510] env[62368]: DEBUG oslo_concurrency.lockutils [req-38450e22-ec4f-4c40-ac78-037346a54bb5 req-5178ecef-0a77-47df-a8b6-bfd9d2557ff6 service nova] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.633648] env[62368]: DEBUG nova.compute.manager [req-38450e22-ec4f-4c40-ac78-037346a54bb5 req-5178ecef-0a77-47df-a8b6-bfd9d2557ff6 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] No waiting events found dispatching network-vif-plugged-58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.633648] env[62368]: WARNING nova.compute.manager [req-38450e22-ec4f-4c40-ac78-037346a54bb5 req-5178ecef-0a77-47df-a8b6-bfd9d2557ff6 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Received unexpected event network-vif-plugged-58072bc6-80bf-4d1f-b471-2ecdd115a8ad for instance with vm_state building and task_state spawning. [ 902.762619] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "5633292c-9a74-4c2d-893b-6823f4a587a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.762952] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.846330] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198723, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.855862] env[62368]: DEBUG oslo_vmware.api [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Task: {'id': task-1198724, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154801} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.858186] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.858394] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 902.858951] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 902.858951] env[62368]: INFO nova.compute.manager [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Took 1.19 seconds to destroy the instance on the hypervisor. [ 902.859062] env[62368]: DEBUG oslo.service.loopingcall [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.860128] env[62368]: DEBUG nova.compute.manager [-] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.860280] env[62368]: DEBUG nova.network.neutron [-] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 902.887985] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198725, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086662} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.892420] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.892420] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42985ed6-a7e3-440d-8900-2f11e4bfa2c7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.897670] env[62368]: DEBUG nova.network.neutron [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Successfully updated port: 58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.918349] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8/14c358a0-434d-472a-a382-b5b0e9bf2bd8.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.922967] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.923139] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.923337] env[62368]: DEBUG nova.network.neutron [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 902.924387] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd0db80e-12f9-4b4b-a454-bfa744034164 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.948300] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 902.948300] env[62368]: value = "task-1198726" [ 902.948300] env[62368]: _type = "Task" [ 902.948300] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.961658] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198726, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.999088] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c4d225-a2d9-4ebd-b8e9-aece072c5ef9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.001423] env[62368]: DEBUG nova.network.neutron [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 903.011522] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ff20d5-6efc-4b2e-a9de-459771f44ba6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.060698] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396b891a-2d9d-4d54-8561-f1ab01909982 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.072331] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c38af3-1adc-453c-9c23-881a0518f569 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.094018] env[62368]: DEBUG nova.compute.provider_tree [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.133786] env[62368]: DEBUG nova.network.neutron [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance_info_cache with network_info: [{"id": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "address": "fa:16:3e:e1:de:4b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be1ff35-0e", "ovs_interfaceid": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.266518] env[62368]: DEBUG nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 903.306373] env[62368]: DEBUG nova.network.neutron [req-c5fc320e-96e0-48fe-bee7-51a9715e1014 req-19f4876e-20f5-41e7-84e0-bb0eb7e6bb93 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Updated VIF entry in instance network info cache for port b8138e3d-417d-4a57-8c35-ac1d61118665. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 903.307161] env[62368]: DEBUG nova.network.neutron [req-c5fc320e-96e0-48fe-bee7-51a9715e1014 req-19f4876e-20f5-41e7-84e0-bb0eb7e6bb93 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Updating instance_info_cache with network_info: [{"id": "b8138e3d-417d-4a57-8c35-ac1d61118665", "address": "fa:16:3e:48:d3:0d", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8138e3d-41", "ovs_interfaceid": "b8138e3d-417d-4a57-8c35-ac1d61118665", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.314130] env[62368]: DEBUG nova.network.neutron [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Updating instance_info_cache with network_info: [{"id": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "address": "fa:16:3e:17:f6:62", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58072bc6-80", "ovs_interfaceid": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.335374] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198723, 'name': CreateVM_Task, 'duration_secs': 0.71057} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.335374] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 903.336026] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.336291] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.336727] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.337155] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1277a510-f946-4662-8ee8-4780d64504c9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.343635] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 903.343635] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ef2b66-e46f-824f-1a1b-1bbfe46bdb4c" [ 903.343635] env[62368]: _type = "Task" [ 903.343635] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.353267] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ef2b66-e46f-824f-1a1b-1bbfe46bdb4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.427470] env[62368]: DEBUG nova.compute.manager [req-0222861e-9ef4-43b1-8b2b-0119bf1a6025 req-59028846-5bca-4af9-81a9-3547332b8a90 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.427470] env[62368]: DEBUG nova.compute.manager [req-0222861e-9ef4-43b1-8b2b-0119bf1a6025 req-59028846-5bca-4af9-81a9-3547332b8a90 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing instance network info cache due to event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 903.427470] env[62368]: DEBUG oslo_concurrency.lockutils [req-0222861e-9ef4-43b1-8b2b-0119bf1a6025 req-59028846-5bca-4af9-81a9-3547332b8a90 service nova] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.427470] env[62368]: DEBUG oslo_concurrency.lockutils [req-0222861e-9ef4-43b1-8b2b-0119bf1a6025 req-59028846-5bca-4af9-81a9-3547332b8a90 service nova] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.427470] env[62368]: DEBUG nova.network.neutron [req-0222861e-9ef4-43b1-8b2b-0119bf1a6025 req-59028846-5bca-4af9-81a9-3547332b8a90 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 903.464021] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198726, 'name': ReconfigVM_Task, 'duration_secs': 0.346496} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.464021] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8/14c358a0-434d-472a-a382-b5b0e9bf2bd8.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.464021] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d9508e4-2f25-4706-9d8e-b4513c909ff9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.472191] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 903.472191] env[62368]: value = "task-1198727" [ 903.472191] env[62368]: _type = "Task" [ 903.472191] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.480819] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198727, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.599899] env[62368]: DEBUG nova.scheduler.client.report [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.640417] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.797411] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.810489] env[62368]: DEBUG oslo_concurrency.lockutils [req-c5fc320e-96e0-48fe-bee7-51a9715e1014 req-19f4876e-20f5-41e7-84e0-bb0eb7e6bb93 service nova] Releasing lock "refresh_cache-ad21bf3a-068e-44a7-b58f-2978a631b972" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.816714] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.817043] env[62368]: DEBUG nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Instance network_info: |[{"id": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "address": "fa:16:3e:17:f6:62", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58072bc6-80", "ovs_interfaceid": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 903.817514] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:f6:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58072bc6-80bf-4d1f-b471-2ecdd115a8ad', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.826292] env[62368]: DEBUG oslo.service.loopingcall [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.826538] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 903.826769] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9b10805-6ae4-4546-bc04-1fe722ffee55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.854953] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ef2b66-e46f-824f-1a1b-1bbfe46bdb4c, 'name': SearchDatastore_Task, 'duration_secs': 0.01016} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.856707] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.856991] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.857261] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.857417] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.857602] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.858836] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.858836] env[62368]: value = "task-1198728" [ 903.858836] env[62368]: _type = "Task" [ 903.858836] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.859097] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23a3639c-ce05-4e13-aeee-d4ef9519bea2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.875456] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198728, 'name': CreateVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.876878] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.877078] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 903.877842] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba514c20-776f-45d6-92e1-399de95f81a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.882415] env[62368]: DEBUG nova.network.neutron [-] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.885999] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 903.885999] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52174270-4ad1-d0d6-b287-6bbf5cdfb5c1" [ 903.885999] env[62368]: _type = "Task" [ 903.885999] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.898970] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52174270-4ad1-d0d6-b287-6bbf5cdfb5c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.983467] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198727, 'name': Rename_Task, 'duration_secs': 0.167062} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.983467] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 903.983467] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c09a1c6-f019-439e-826a-62e757e4057f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.990596] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 903.990596] env[62368]: value = "task-1198729" [ 903.990596] env[62368]: _type = "Task" [ 903.990596] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.000457] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.110025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.110025] env[62368]: DEBUG nova.compute.manager [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.112049] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.293s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.272875] env[62368]: DEBUG nova.network.neutron [req-0222861e-9ef4-43b1-8b2b-0119bf1a6025 req-59028846-5bca-4af9-81a9-3547332b8a90 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updated VIF entry in instance network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 904.273391] env[62368]: DEBUG nova.network.neutron [req-0222861e-9ef4-43b1-8b2b-0119bf1a6025 req-59028846-5bca-4af9-81a9-3547332b8a90 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.373887] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198728, 'name': CreateVM_Task, 'duration_secs': 0.452598} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.374088] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 904.374836] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.375096] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.375458] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.376600] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12c698e3-01fd-4c7a-9cf2-bbb458ae87fb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.382929] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 904.382929] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52905257-0a11-81d0-83d2-052d1e800647" [ 904.382929] env[62368]: _type = "Task" [ 904.382929] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.386317] env[62368]: INFO nova.compute.manager [-] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Took 1.53 seconds to deallocate network for instance. [ 904.398276] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52905257-0a11-81d0-83d2-052d1e800647, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.405858] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52174270-4ad1-d0d6-b287-6bbf5cdfb5c1, 'name': SearchDatastore_Task, 'duration_secs': 0.011771} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.406750] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e287f4e-dd9b-414f-8e39-03b43a634a1b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.413374] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 904.413374] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f5d5e3-1ff2-1860-0a4e-8880d85d8378" [ 904.413374] env[62368]: _type = "Task" [ 904.413374] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.423565] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f5d5e3-1ff2-1860-0a4e-8880d85d8378, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.501011] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198729, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.624614] env[62368]: DEBUG nova.compute.utils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.626192] env[62368]: DEBUG nova.compute.manager [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Not allocating networking since 'none' was specified. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 904.760255] env[62368]: DEBUG nova.compute.manager [req-f24a814d-69af-43f2-a6dd-ce78e1324f32 req-f9985a7d-c63e-48ad-b699-9f3a519a2178 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Received event network-changed-58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.760896] env[62368]: DEBUG nova.compute.manager [req-f24a814d-69af-43f2-a6dd-ce78e1324f32 req-f9985a7d-c63e-48ad-b699-9f3a519a2178 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Refreshing instance network info cache due to event network-changed-58072bc6-80bf-4d1f-b471-2ecdd115a8ad. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.760896] env[62368]: DEBUG oslo_concurrency.lockutils [req-f24a814d-69af-43f2-a6dd-ce78e1324f32 req-f9985a7d-c63e-48ad-b699-9f3a519a2178 service nova] Acquiring lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.761020] env[62368]: DEBUG oslo_concurrency.lockutils [req-f24a814d-69af-43f2-a6dd-ce78e1324f32 req-f9985a7d-c63e-48ad-b699-9f3a519a2178 service nova] Acquired lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.761216] env[62368]: DEBUG nova.network.neutron [req-f24a814d-69af-43f2-a6dd-ce78e1324f32 req-f9985a7d-c63e-48ad-b699-9f3a519a2178 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Refreshing network info cache for port 58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.776244] env[62368]: DEBUG oslo_concurrency.lockutils [req-0222861e-9ef4-43b1-8b2b-0119bf1a6025 req-59028846-5bca-4af9-81a9-3547332b8a90 service nova] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.896708] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52905257-0a11-81d0-83d2-052d1e800647, 'name': SearchDatastore_Task, 'duration_secs': 0.011213} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.897390] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.897680] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.897907] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.898831] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.924565] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f5d5e3-1ff2-1860-0a4e-8880d85d8378, 'name': SearchDatastore_Task, 'duration_secs': 0.010676} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.924859] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.925178] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] ad21bf3a-068e-44a7-b58f-2978a631b972/ad21bf3a-068e-44a7-b58f-2978a631b972.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 904.925480] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.925711] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.925940] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21ed612a-87bb-45c9-9ef7-5922652b935a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.928115] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-354b397f-aeb5-4d3b-a808-e22a0461b273 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.937466] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 904.937466] env[62368]: value = "task-1198730" [ 904.937466] env[62368]: _type = "Task" [ 904.937466] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.943308] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.943639] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 904.944762] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd30f553-3c78-4338-84a9-275942777725 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.950471] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198730, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.954248] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 904.954248] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5237d2d3-f45a-77d8-39a7-ab27848d8a85" [ 904.954248] env[62368]: _type = "Task" [ 904.954248] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.962677] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5237d2d3-f45a-77d8-39a7-ab27848d8a85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.999557] env[62368]: DEBUG oslo_vmware.api [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198729, 'name': PowerOnVM_Task, 'duration_secs': 0.54061} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.000103] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 905.000975] env[62368]: INFO nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Took 8.69 seconds to spawn the instance on the hypervisor. [ 905.001249] env[62368]: DEBUG nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.002064] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77888d1c-c146-4b65-aa3d-de17f67037a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.128114] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Applying migration context for instance 13765305-2e55-4ee8-9a6f-4ae5ee724367 as it has an incoming, in-progress migration a4a55aca-f211-40de-8c6e-9835ad122ae2. Migration status is confirming {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 905.128576] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Applying migration context for instance 5af32858-dc9a-4380-827a-daf384c29bed as it has an incoming, in-progress migration db16e018-9b81-40e4-8b6b-0ac690c9f100. Migration status is migrating {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 905.131872] env[62368]: INFO nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating resource usage from migration a4a55aca-f211-40de-8c6e-9835ad122ae2 [ 905.132129] env[62368]: INFO nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating resource usage from migration db16e018-9b81-40e4-8b6b-0ac690c9f100 [ 905.135529] env[62368]: DEBUG nova.compute.manager [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.158609] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a59f93-50cd-461a-9e5f-0f35e69f8361 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.165084] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b79f0e79-9e3a-47c7-9949-8743601ec6c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.165322] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance eea21546-fbbf-4440-829c-8583c4ccabb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.165452] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 58157ab1-80a4-427c-812b-f6fde1f8db68 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.165633] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 2a46c954-449e-4d62-be80-add1040ed4c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.165802] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b87094da-6258-469e-ab37-5557955ad3a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.165962] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 69ddb565-6c79-44e5-a7d1-d339ab426fae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.166466] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 55eecf3d-501c-490d-bddd-0211fd082841 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.166629] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Migration a4a55aca-f211-40de-8c6e-9835ad122ae2 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 905.166773] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 13765305-2e55-4ee8-9a6f-4ae5ee724367 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.166994] env[62368]: WARNING nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 207de202-ca50-4811-84c5-a0d4454d29f7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 905.167282] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a5cbadbd-20dd-4514-8867-20243af5db0c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.167440] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 14c358a0-434d-472a-a382-b5b0e9bf2bd8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.167574] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance ad21bf3a-068e-44a7-b58f-2978a631b972 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.167745] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.167877] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Migration db16e018-9b81-40e4-8b6b-0ac690c9f100 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 905.168055] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 5af32858-dc9a-4380-827a-daf384c29bed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.168194] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 905.188123] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance '5af32858-dc9a-4380-827a-daf384c29bed' progress to 0 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 905.192212] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 3e09be20-ce79-4bfa-bf84-2877f4e534e6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 905.449555] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198730, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.478795] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5237d2d3-f45a-77d8-39a7-ab27848d8a85, 'name': SearchDatastore_Task, 'duration_secs': 0.021423} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.479872] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f4b2869-19d8-4514-8869-b2ff521dbb5f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.489535] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 905.489535] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e65645-da3a-6721-0dee-4a6d509110e9" [ 905.489535] env[62368]: _type = "Task" [ 905.489535] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.498120] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e65645-da3a-6721-0dee-4a6d509110e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.521947] env[62368]: INFO nova.compute.manager [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Took 20.54 seconds to build instance. [ 905.572238] env[62368]: DEBUG nova.compute.manager [req-de0feefe-eb09-447b-85ac-ad0680c48288 req-592627de-0635-4457-a56c-811fdc767b47 service nova] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Received event network-vif-deleted-52cba1c6-211a-412d-9265-34ce316a95a3 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.621088] env[62368]: DEBUG nova.network.neutron [req-f24a814d-69af-43f2-a6dd-ce78e1324f32 req-f9985a7d-c63e-48ad-b699-9f3a519a2178 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Updated VIF entry in instance network info cache for port 58072bc6-80bf-4d1f-b471-2ecdd115a8ad. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.621088] env[62368]: DEBUG nova.network.neutron [req-f24a814d-69af-43f2-a6dd-ce78e1324f32 req-f9985a7d-c63e-48ad-b699-9f3a519a2178 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Updating instance_info_cache with network_info: [{"id": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "address": "fa:16:3e:17:f6:62", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58072bc6-80", "ovs_interfaceid": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.698022] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.698022] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 5633292c-9a74-4c2d-893b-6823f4a587a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 905.698022] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 905.698022] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3712MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 905.702721] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcf2176c-daab-4400-9cb9-14869eadf4f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.712600] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 905.712600] env[62368]: value = "task-1198731" [ 905.712600] env[62368]: _type = "Task" [ 905.712600] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.724795] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.955518] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198730, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5241} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.956021] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] ad21bf3a-068e-44a7-b58f-2978a631b972/ad21bf3a-068e-44a7-b58f-2978a631b972.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 905.956389] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.959100] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1cfc2dcb-a147-4bcd-ab95-46d90ea96edd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.967360] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 905.967360] env[62368]: value = "task-1198732" [ 905.967360] env[62368]: _type = "Task" [ 905.967360] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.980516] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.986218] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c056cb-8ddb-4dd1-8194-601d1e6c1ac3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.999059] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e65645-da3a-6721-0dee-4a6d509110e9, 'name': SearchDatastore_Task, 'duration_secs': 0.010186} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.000847] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.001153] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e/3209ccfd-34e7-4c95-ba24-3d86fcdbae6e.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 906.001527] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03ceba06-bd6b-4123-883a-1ba49dc39ec8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.004378] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061ef57f-be8a-4797-ae00-650d9cffdc65 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.039080] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c0e145dc-abd9-427e-b7dd-11dc8617df08 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.075s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.041779] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53e6d4e-1e8b-429e-b2a4-128f74c95402 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.044809] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 906.044809] env[62368]: value = "task-1198733" [ 906.044809] env[62368]: _type = "Task" [ 906.044809] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.053715] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807ec002-4048-4229-acf7-11647f340ead {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.065702] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198733, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.076933] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.123181] env[62368]: DEBUG oslo_concurrency.lockutils [req-f24a814d-69af-43f2-a6dd-ce78e1324f32 req-f9985a7d-c63e-48ad-b699-9f3a519a2178 service nova] Releasing lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.149302] env[62368]: DEBUG nova.compute.manager [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.178646] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.178952] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.179163] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.179377] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.179532] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.179714] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.179905] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.180115] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.180349] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.180535] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.180719] env[62368]: DEBUG nova.virt.hardware [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.181727] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48aaca7c-55b4-4c63-b48e-3b8b81c879db {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.191952] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986bddc1-bc46-4f80-8e9d-a27799733813 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.209490] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.215291] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Creating folder: Project (237a0ace810a4893b73f6ca92adbfa8b). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 906.215664] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4a12e9f-404a-44fa-9e58-35a12ea3b030 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.230395] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198731, 'name': PowerOffVM_Task, 'duration_secs': 0.423151} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.230606] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 906.230682] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance '5af32858-dc9a-4380-827a-daf384c29bed' progress to 17 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 906.236366] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Created folder: Project (237a0ace810a4893b73f6ca92adbfa8b) in parent group-v259706. [ 906.236601] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Creating folder: Instances. Parent ref: group-v259848. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 906.237176] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-13f05a6a-cc33-404c-a891-af85a2ac6095 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.250037] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Created folder: Instances in parent group-v259848. [ 906.250300] env[62368]: DEBUG oslo.service.loopingcall [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.250562] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 906.250807] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-100012dc-b629-4ba7-a8d1-e7b33f769816 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.271875] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.271875] env[62368]: value = "task-1198736" [ 906.271875] env[62368]: _type = "Task" [ 906.271875] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.281883] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198736, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.481525] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074491} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.481525] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.481525] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db527ebe-9e9d-458d-8862-880db0de6ab7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.507401] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] ad21bf3a-068e-44a7-b58f-2978a631b972/ad21bf3a-068e-44a7-b58f-2978a631b972.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.507881] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-888575c5-6804-4ce1-8d9b-3e69f6310722 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.530987] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 906.530987] env[62368]: value = "task-1198737" [ 906.530987] env[62368]: _type = "Task" [ 906.530987] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.546587] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198737, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.555464] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198733, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490141} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.555740] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e/3209ccfd-34e7-4c95-ba24-3d86fcdbae6e.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 906.556035] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.556449] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c84f12d-438f-4fbd-b25c-5991d51fc274 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.563345] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 906.563345] env[62368]: value = "task-1198738" [ 906.563345] env[62368]: _type = "Task" [ 906.563345] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.567295] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.567522] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.567702] env[62368]: DEBUG nova.compute.manager [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.568783] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a82e77b-7bcc-4473-bdfd-a2da5662cd74 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.578062] env[62368]: DEBUG nova.compute.manager [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62368) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 906.579121] env[62368]: DEBUG nova.objects.instance [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lazy-loading 'flavor' on Instance uuid 14c358a0-434d-472a-a382-b5b0e9bf2bd8 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.585019] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.588879] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198738, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.741080] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.741080] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.741080] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.741080] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.741552] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.741880] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.742281] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.743927] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.743927] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.743927] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.743927] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.749380] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6150f849-7753-4ff0-81f6-b7c4cf63efe9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.768048] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 906.768048] env[62368]: value = "task-1198739" [ 906.768048] env[62368]: _type = "Task" [ 906.768048] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.779822] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.785665] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198736, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.043803] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198737, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.073575] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198738, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.219226} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.073782] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.074635] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78a99f7-a109-4387-a1e9-4140613817e6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.100711] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e/3209ccfd-34e7-4c95-ba24-3d86fcdbae6e.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.101375] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 907.101981] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 907.102192] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.990s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.102400] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f009827f-5ccf-45e7-9f45-750b97b09196 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.117301] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3d053ec-f631-4e1c-a01b-472d4c5f1838 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.118960] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.958s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.119214] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.121412] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.171s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.123151] env[62368]: INFO nova.compute.claims [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.134490] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 907.134490] env[62368]: value = "task-1198741" [ 907.134490] env[62368]: _type = "Task" [ 907.134490] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.135475] env[62368]: DEBUG oslo_vmware.api [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 907.135475] env[62368]: value = "task-1198740" [ 907.135475] env[62368]: _type = "Task" [ 907.135475] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.150091] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198741, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.153369] env[62368]: DEBUG oslo_vmware.api [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.163245] env[62368]: INFO nova.scheduler.client.report [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Deleted allocations for instance 207de202-ca50-4811-84c5-a0d4454d29f7 [ 907.285877] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198739, 'name': ReconfigVM_Task, 'duration_secs': 0.468448} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.289814] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance '5af32858-dc9a-4380-827a-daf384c29bed' progress to 33 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 907.293286] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198736, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.542752] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198737, 'name': ReconfigVM_Task, 'duration_secs': 0.900186} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.542955] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Reconfigured VM instance instance-00000055 to attach disk [datastore2] ad21bf3a-068e-44a7-b58f-2978a631b972/ad21bf3a-068e-44a7-b58f-2978a631b972.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.543616] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b12dbafa-1936-4b7f-96c5-7d11dd31f84c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.552130] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 907.552130] env[62368]: value = "task-1198742" [ 907.552130] env[62368]: _type = "Task" [ 907.552130] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.561184] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198742, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.649730] env[62368]: DEBUG oslo_vmware.api [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198740, 'name': PowerOffVM_Task, 'duration_secs': 0.259741} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.653451] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 907.653596] env[62368]: DEBUG nova.compute.manager [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.653892] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198741, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.654875] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fee7992-4d97-465d-92a1-d7415244eeb6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.674688] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f5f048d6-e5ac-4f39-addb-f15f1903b094 tempest-ImagesOneServerNegativeTestJSON-1573857276 tempest-ImagesOneServerNegativeTestJSON-1573857276-project-member] Lock "207de202-ca50-4811-84c5-a0d4454d29f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.677s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.785035] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198736, 'name': CreateVM_Task, 'duration_secs': 1.489791} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.785035] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 907.785734] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.785734] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.785957] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.786229] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e85171b-326b-4444-9eef-e85e8224a36b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.791466] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 907.791466] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52026dce-e3c1-9b41-89ac-b73beccca880" [ 907.791466] env[62368]: _type = "Task" [ 907.791466] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.796144] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.796374] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.796538] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.796730] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.796939] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.797028] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.797238] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.797415] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.797585] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.797750] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.797925] env[62368]: DEBUG nova.virt.hardware [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.803126] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Reconfiguring VM instance instance-00000051 to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 907.803392] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96dbdf71-2289-449c-bd3d-623601fcd7f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.816774] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "eea21546-fbbf-4440-829c-8583c4ccabb6" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.817017] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.817233] env[62368]: INFO nova.compute.manager [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Shelving [ 907.826021] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52026dce-e3c1-9b41-89ac-b73beccca880, 'name': SearchDatastore_Task, 'duration_secs': 0.010473} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.826021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.826021] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.826021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.826021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.826021] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.826021] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b02371af-afac-4aba-a6fe-f31a501a8300 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.828945] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 907.828945] env[62368]: value = "task-1198743" [ 907.828945] env[62368]: _type = "Task" [ 907.828945] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.837319] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.837657] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 907.843652] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-732f35d7-6798-4ca0-afd1-6c0200220736 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.848893] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198743, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.853707] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 907.853707] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521f8cd4-6553-513e-2992-a92768e18231" [ 907.853707] env[62368]: _type = "Task" [ 907.853707] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.869223] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521f8cd4-6553-513e-2992-a92768e18231, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.869982] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04547bc8-5847-476c-9caa-00a43cf423cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.875934] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 907.875934] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5201a8c0-56f8-0760-fa74-bee6fe38e35c" [ 907.875934] env[62368]: _type = "Task" [ 907.875934] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.884222] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5201a8c0-56f8-0760-fa74-bee6fe38e35c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.063520] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198742, 'name': Rename_Task, 'duration_secs': 0.176938} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.063852] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 908.064140] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9ee2fad-7178-4f44-8229-62a69941c53c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.072425] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 908.072425] env[62368]: value = "task-1198744" [ 908.072425] env[62368]: _type = "Task" [ 908.072425] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.082091] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198744, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.151595] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198741, 'name': ReconfigVM_Task, 'duration_secs': 0.541508} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.152239] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e/3209ccfd-34e7-4c95-ba24-3d86fcdbae6e.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.153409] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ecc7e47-880e-4f49-ab73-6bfd30749c89 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.170346] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 908.170346] env[62368]: value = "task-1198745" [ 908.170346] env[62368]: _type = "Task" [ 908.170346] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.171217] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c615a653-455a-4e01-8c7f-c97812856954 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.604s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.194762] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198745, 'name': Rename_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.330909] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 908.334789] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-326da6ad-547f-4eeb-91e1-35be56d34784 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.344823] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198743, 'name': ReconfigVM_Task, 'duration_secs': 0.310291} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.346237] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Reconfigured VM instance instance-00000051 to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 908.346609] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 908.346609] env[62368]: value = "task-1198746" [ 908.346609] env[62368]: _type = "Task" [ 908.346609] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.347414] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a50b02-dc6f-4885-a200-c751ee1a6cef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.361758] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.382752] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 5af32858-dc9a-4380-827a-daf384c29bed/5af32858-dc9a-4380-827a-daf384c29bed.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.386019] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-def17c76-0515-46ed-b68b-cae625be6fb7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.411858] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5201a8c0-56f8-0760-fa74-bee6fe38e35c, 'name': SearchDatastore_Task, 'duration_secs': 0.010409} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.413419] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.413735] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc/7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 908.414088] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 908.414088] env[62368]: value = "task-1198747" [ 908.414088] env[62368]: _type = "Task" [ 908.414088] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.416907] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-086d1446-0503-43fa-8555-48ba4aa4e7a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.429174] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198747, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.430870] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 908.430870] env[62368]: value = "task-1198748" [ 908.430870] env[62368]: _type = "Task" [ 908.430870] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.444331] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198748, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.463641] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ee6a5b-e779-4d99-b243-1ea100fbb9fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.472510] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffb6250-6c3a-4ab7-a053-cb50228b6d72 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.510393] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94407258-11b1-4699-bbcc-57bad3985c72 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.519473] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d03d070-2c67-4d0e-b9cc-cc307c61d50b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.535748] env[62368]: DEBUG nova.compute.provider_tree [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 908.583977] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198744, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.689494] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198745, 'name': Rename_Task, 'duration_secs': 0.263796} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.689883] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 908.690330] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6846c149-9b6c-4307-adc6-b58ab63ae03d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.700902] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 908.700902] env[62368]: value = "task-1198749" [ 908.700902] env[62368]: _type = "Task" [ 908.700902] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.719055] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198749, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.861393] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198746, 'name': PowerOffVM_Task, 'duration_secs': 0.23371} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.861747] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 908.862691] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dafb62e5-2f20-4c21-bea6-7991ef9eabd5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.887258] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca27d712-fe41-4068-bb8c-55434a292239 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.932466] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198747, 'name': ReconfigVM_Task, 'duration_secs': 0.444494} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.932812] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 5af32858-dc9a-4380-827a-daf384c29bed/5af32858-dc9a-4380-827a-daf384c29bed.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.933188] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance '5af32858-dc9a-4380-827a-daf384c29bed' progress to 50 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 908.946727] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198748, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505526} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.947144] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc/7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 908.947621] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.948178] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5f2a48f-537d-4806-9a31-925401582fe6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.956436] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 908.956436] env[62368]: value = "task-1198750" [ 908.956436] env[62368]: _type = "Task" [ 908.956436] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.965069] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198750, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.056403] env[62368]: ERROR nova.scheduler.client.report [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [req-c01b469c-44a8-466e-8a4e-1947451c82f6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c01b469c-44a8-466e-8a4e-1947451c82f6"}]} [ 909.072045] env[62368]: DEBUG nova.scheduler.client.report [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 909.083861] env[62368]: DEBUG oslo_vmware.api [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198744, 'name': PowerOnVM_Task, 'duration_secs': 0.675373} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.084151] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 909.084362] env[62368]: INFO nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Took 10.31 seconds to spawn the instance on the hypervisor. [ 909.084541] env[62368]: DEBUG nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.085369] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a92a9d-4955-469e-95e9-546c4ed0656e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.098497] env[62368]: DEBUG nova.scheduler.client.report [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 909.098497] env[62368]: DEBUG nova.compute.provider_tree [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 909.109290] env[62368]: DEBUG nova.scheduler.client.report [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 909.134317] env[62368]: DEBUG nova.scheduler.client.report [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 909.213750] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198749, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.405852] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 909.406479] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-fc78bb37-52d3-4bb4-aac0-f01256dc6f74 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.413139] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e90fbfd-b28a-4e52-a5c4-3910d04d1c38 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.417939] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 909.417939] env[62368]: value = "task-1198751" [ 909.417939] env[62368]: _type = "Task" [ 909.417939] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.426335] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e20bac-b860-4a39-b28a-22f3cacd9cfd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.432414] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198751, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.462541] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95056964-16e5-4217-a066-d800222e3eed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.468602] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c14ec43-fea5-4680-b328-06c46b7b2c3f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.481151] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198750, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.213192} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.494516] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.495947] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b5521e-8725-4ce8-8b26-949a84a46899 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.500759] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8459fa-4f01-437a-a90d-e351dfdef4e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.503549] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddcedd7-2dd7-449b-a648-0de02f711eca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.530038] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance '5af32858-dc9a-4380-827a-daf384c29bed' progress to 67 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 909.545304] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc/7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.551857] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-994d85f9-eaf5-4fbb-af3b-a9bff39bf4e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.566234] env[62368]: DEBUG nova.compute.provider_tree [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.574340] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 909.574340] env[62368]: value = "task-1198752" [ 909.574340] env[62368]: _type = "Task" [ 909.574340] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.585803] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198752, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.608022] env[62368]: INFO nova.compute.manager [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Took 20.31 seconds to build instance. [ 909.712426] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198749, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.775026] env[62368]: INFO nova.compute.manager [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Rebuilding instance [ 909.830887] env[62368]: DEBUG nova.compute.manager [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.831868] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e80edf1-760c-47e4-820e-271850980ccb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.889186] env[62368]: DEBUG oslo_concurrency.lockutils [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.889186] env[62368]: DEBUG oslo_concurrency.lockutils [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.931230] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198751, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.069097] env[62368]: DEBUG nova.scheduler.client.report [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.084851] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198752, 'name': ReconfigVM_Task, 'duration_secs': 0.318127} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.085370] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc/7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.086370] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56aae2ea-4adb-4949-b2c0-353c9b4c49d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.093727] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 910.093727] env[62368]: value = "task-1198753" [ 910.093727] env[62368]: _type = "Task" [ 910.093727] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.104638] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198753, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.109676] env[62368]: DEBUG oslo_concurrency.lockutils [None req-66d961d8-59eb-4704-9c07-719301e00f3f tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "ad21bf3a-068e-44a7-b58f-2978a631b972" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.821s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.129810] env[62368]: DEBUG nova.network.neutron [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Port 8be1ff35-0e69-42f1-a582-16bf1f496a2d binding to destination host cpu-1 is already ACTIVE {{(pid=62368) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 910.214278] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198749, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.267608] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "ad21bf3a-068e-44a7-b58f-2978a631b972" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.267942] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "ad21bf3a-068e-44a7-b58f-2978a631b972" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.268270] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "ad21bf3a-068e-44a7-b58f-2978a631b972-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.268517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "ad21bf3a-068e-44a7-b58f-2978a631b972-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.268769] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "ad21bf3a-068e-44a7-b58f-2978a631b972-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.271486] env[62368]: INFO nova.compute.manager [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Terminating instance [ 910.273392] env[62368]: DEBUG nova.compute.manager [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.273610] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 910.274552] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10989342-0ea4-455b-a376-ce71a9189f9b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.288921] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.290125] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd7ba3e5-0086-4851-8cd5-f3881c892a8c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.299393] env[62368]: DEBUG oslo_vmware.api [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 910.299393] env[62368]: value = "task-1198754" [ 910.299393] env[62368]: _type = "Task" [ 910.299393] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.311022] env[62368]: DEBUG oslo_vmware.api [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198754, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.349016] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.349016] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ad94947-afce-4700-a3be-d2bde70c0b7d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.355190] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 910.355190] env[62368]: value = "task-1198755" [ 910.355190] env[62368]: _type = "Task" [ 910.355190] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.373312] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 910.373532] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 910.375150] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03751e76-eb08-40bb-a7ae-038843d060f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.383564] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.383736] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68b4b093-6d38-4ce2-bb95-f1c7e6a1c8fb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.394233] env[62368]: INFO nova.compute.manager [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Detaching volume 11c38512-c804-4553-9c51-d6a79caa8f45 [ 910.429310] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198751, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.445583] env[62368]: INFO nova.virt.block_device [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Attempting to driver detach volume 11c38512-c804-4553-9c51-d6a79caa8f45 from mountpoint /dev/sdb [ 910.446582] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 910.446582] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259834', 'volume_id': '11c38512-c804-4553-9c51-d6a79caa8f45', 'name': 'volume-11c38512-c804-4553-9c51-d6a79caa8f45', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '58157ab1-80a4-427c-812b-f6fde1f8db68', 'attached_at': '', 'detached_at': '', 'volume_id': '11c38512-c804-4553-9c51-d6a79caa8f45', 'serial': '11c38512-c804-4553-9c51-d6a79caa8f45'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 910.447991] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e63cb01-bee6-4df4-8fd1-dd6ca1d62f69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.476676] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbcffedd-e0b9-46d4-ab14-f4b3d85b976f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.487994] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d641e80-d00a-4414-b395-045665f64855 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.512335] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56aa7149-76e1-4c39-a183-4340e53ebe33 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.530648] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] The volume has not been displaced from its original location: [datastore1] volume-11c38512-c804-4553-9c51-d6a79caa8f45/volume-11c38512-c804-4553-9c51-d6a79caa8f45.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 910.536618] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Reconfiguring VM instance instance-00000041 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 910.538220] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be3fa70d-4d03-4776-9020-910a0b4e9a3c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.554597] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.556266] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.556266] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleting the datastore file [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.556266] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-defdb187-e38d-4d0f-8bcd-8968406e641f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.565140] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 910.565140] env[62368]: value = "task-1198757" [ 910.565140] env[62368]: _type = "Task" [ 910.565140] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.566474] env[62368]: DEBUG oslo_vmware.api [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 910.566474] env[62368]: value = "task-1198758" [ 910.566474] env[62368]: _type = "Task" [ 910.566474] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.574688] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.453s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.575324] env[62368]: DEBUG nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.581992] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.456s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.583781] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198757, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.589854] env[62368]: DEBUG oslo_vmware.api [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198758, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.603512] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198753, 'name': Rename_Task, 'duration_secs': 0.155733} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.604453] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 910.604748] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90330b20-2c58-45be-a903-51bd77f2361e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.613543] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 910.613543] env[62368]: value = "task-1198759" [ 910.613543] env[62368]: _type = "Task" [ 910.613543] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.623221] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198759, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.716435] env[62368]: DEBUG oslo_vmware.api [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198749, 'name': PowerOnVM_Task, 'duration_secs': 1.836291} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.716435] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 910.716435] env[62368]: INFO nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Took 9.46 seconds to spawn the instance on the hypervisor. [ 910.716435] env[62368]: DEBUG nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.717258] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884d91e4-43b1-4ab0-8761-bd9393353ea3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.810442] env[62368]: DEBUG oslo_vmware.api [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198754, 'name': PowerOffVM_Task, 'duration_secs': 0.345171} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.810750] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.810884] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.811181] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6823ef6e-cd40-4bc1-9403-1592525743e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.901688] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.901975] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.902235] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleting the datastore file [datastore2] ad21bf3a-068e-44a7-b58f-2978a631b972 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.902537] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0f75f6f-4bea-43f1-b93c-93d91f49cb47 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.910869] env[62368]: DEBUG oslo_vmware.api [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 910.910869] env[62368]: value = "task-1198761" [ 910.910869] env[62368]: _type = "Task" [ 910.910869] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.921520] env[62368]: DEBUG oslo_vmware.api [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198761, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.937459] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198751, 'name': CreateSnapshot_Task, 'duration_secs': 1.195844} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.937780] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 910.938609] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5029e3ed-45b9-44a2-82df-c7b8516e184f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.078615] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198757, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20819} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.082059] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.082059] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.082255] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 911.084826] env[62368]: DEBUG oslo_vmware.api [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198758, 'name': ReconfigVM_Task, 'duration_secs': 0.330641} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.088727] env[62368]: DEBUG nova.compute.utils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.090084] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Reconfigured VM instance instance-00000041 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 911.096165] env[62368]: DEBUG nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 911.096165] env[62368]: DEBUG nova.network.neutron [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 911.097757] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c16e3b72-4ff9-409d-87d2-e5f80c0b879f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.120135] env[62368]: DEBUG oslo_vmware.api [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 911.120135] env[62368]: value = "task-1198762" [ 911.120135] env[62368]: _type = "Task" [ 911.120135] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.130659] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198759, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.149117] env[62368]: DEBUG oslo_vmware.api [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198762, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.160596] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "5af32858-dc9a-4380-827a-daf384c29bed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.160823] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.161050] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.179524] env[62368]: DEBUG nova.policy [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32da75d294d548eb8c5f75ae18e79c0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3940819fc14a4c628aacd8820efe084e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.245392] env[62368]: INFO nova.compute.manager [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Took 20.86 seconds to build instance. [ 911.381705] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9313e2f-c41d-4f24-a106-7994af07cd75 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.389814] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ca61cb-5d5f-40dd-a588-56f4882a20dd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.423437] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c7928a-867a-4b72-a2f1-044d25852afd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.431362] env[62368]: DEBUG oslo_vmware.api [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198761, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210011} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.433495] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.433691] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.433872] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 911.434083] env[62368]: INFO nova.compute.manager [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Took 1.16 seconds to destroy the instance on the hypervisor. [ 911.434333] env[62368]: DEBUG oslo.service.loopingcall [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.434581] env[62368]: DEBUG nova.compute.manager [-] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.434675] env[62368]: DEBUG nova.network.neutron [-] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.437265] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9287e7-f882-4747-bcfb-b664f06aee8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.457353] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 911.457816] env[62368]: DEBUG nova.compute.provider_tree [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.458968] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-96f2a196-3c1a-43ee-969f-413a631c0a8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.468347] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 911.468347] env[62368]: value = "task-1198763" [ 911.468347] env[62368]: _type = "Task" [ 911.468347] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.478625] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198763, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.576629] env[62368]: DEBUG nova.network.neutron [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Successfully created port: 561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.596325] env[62368]: DEBUG nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.627430] env[62368]: DEBUG oslo_vmware.api [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198759, 'name': PowerOnVM_Task, 'duration_secs': 0.559718} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.632738] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 911.633070] env[62368]: INFO nova.compute.manager [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Took 5.48 seconds to spawn the instance on the hypervisor. [ 911.633864] env[62368]: DEBUG nova.compute.manager [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.634842] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baac2b34-0b9d-4830-9747-d340d7f760ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.651565] env[62368]: DEBUG oslo_vmware.api [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198762, 'name': ReconfigVM_Task, 'duration_secs': 0.158695} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.655624] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259834', 'volume_id': '11c38512-c804-4553-9c51-d6a79caa8f45', 'name': 'volume-11c38512-c804-4553-9c51-d6a79caa8f45', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '58157ab1-80a4-427c-812b-f6fde1f8db68', 'attached_at': '', 'detached_at': '', 'volume_id': '11c38512-c804-4553-9c51-d6a79caa8f45', 'serial': '11c38512-c804-4553-9c51-d6a79caa8f45'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 911.750139] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cfddf3fe-901b-438a-9834-6bd161ca6cda tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.370s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.949686] env[62368]: DEBUG nova.compute.manager [req-2745ca6b-a51d-4381-afd4-11cca8503cde req-92bb9145-a319-4c61-8b7d-70ff17eeea59 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Received event network-vif-deleted-b8138e3d-417d-4a57-8c35-ac1d61118665 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.949898] env[62368]: INFO nova.compute.manager [req-2745ca6b-a51d-4381-afd4-11cca8503cde req-92bb9145-a319-4c61-8b7d-70ff17eeea59 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Neutron deleted interface b8138e3d-417d-4a57-8c35-ac1d61118665; detaching it from the instance and deleting it from the info cache [ 911.950091] env[62368]: DEBUG nova.network.neutron [req-2745ca6b-a51d-4381-afd4-11cca8503cde req-92bb9145-a319-4c61-8b7d-70ff17eeea59 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.963055] env[62368]: DEBUG nova.scheduler.client.report [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.979384] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198763, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.127994] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.128310] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.128486] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.128683] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.128837] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.128988] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.129235] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.129444] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.129616] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.129777] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.133025] env[62368]: DEBUG nova.virt.hardware [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.133025] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec67a08-a23c-4b85-a1fe-4beec21e90ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.141927] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc62052-7124-4908-83b2-47e34ebe58e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.174814] env[62368]: INFO nova.compute.manager [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Took 15.74 seconds to build instance. [ 912.179263] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:57:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41e7d4b7-38bf-4270-812a-ac72f69e40d6', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.190472] env[62368]: DEBUG oslo.service.loopingcall [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.191344] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 912.191954] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ddae9864-cbfd-4461-87cf-8ccf4314714b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.215829] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.215829] env[62368]: value = "task-1198764" [ 912.215829] env[62368]: _type = "Task" [ 912.215829] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.221396] env[62368]: DEBUG nova.objects.instance [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.225967] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198764, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.226877] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.227074] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.227260] env[62368]: DEBUG nova.network.neutron [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 912.299453] env[62368]: DEBUG nova.network.neutron [-] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.455367] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5731e438-f479-43e9-ab7a-b774ce958e7e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.466908] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0a6c6d-fbf0-419e-beb6-dd56355f597e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.493997] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198763, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.508788] env[62368]: DEBUG nova.compute.manager [req-2745ca6b-a51d-4381-afd4-11cca8503cde req-92bb9145-a319-4c61-8b7d-70ff17eeea59 service nova] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Detach interface failed, port_id=b8138e3d-417d-4a57-8c35-ac1d61118665, reason: Instance ad21bf3a-068e-44a7-b58f-2978a631b972 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 912.615155] env[62368]: DEBUG nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.651268] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.651482] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.651684] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.651922] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.652104] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.652342] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.652668] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.652938] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.653135] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.653343] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.653588] env[62368]: DEBUG nova.virt.hardware [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.654628] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eda08e8-259d-4430-a862-9771abbb723e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.665866] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b547dc9d-24ff-4d7b-8792-456e97f506ff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.691960] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f8f59f56-519c-4a63-bc01-58990521367a tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.268s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.738127] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198764, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.760511] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.760511] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.760511] env[62368]: INFO nova.compute.manager [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Shelving [ 912.803624] env[62368]: INFO nova.compute.manager [-] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Took 1.37 seconds to deallocate network for instance. [ 912.825017] env[62368]: INFO nova.compute.manager [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Rebuilding instance [ 912.876485] env[62368]: DEBUG nova.compute.manager [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.880019] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e204733-ec14-4d37-8064-9ed1e953516c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.999099] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.413s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.999412] env[62368]: DEBUG nova.compute.manager [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62368) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 913.005896] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.209s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.007529] env[62368]: INFO nova.compute.claims [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.018158] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198763, 'name': CloneVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.174907] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.230262] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198764, 'name': CreateVM_Task, 'duration_secs': 0.653134} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.230488] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 913.231209] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.231435] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.231815] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.232227] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce201f61-115c-4a69-af59-ae38ea3bffab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.237901] env[62368]: DEBUG oslo_concurrency.lockutils [None req-af171632-5e09-46ed-a8ef-1344ca187f44 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.349s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.240916] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.066s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.241143] env[62368]: DEBUG nova.compute.manager [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.241875] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 913.241875] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e722e8-5b8f-c75f-f168-b668c184d851" [ 913.241875] env[62368]: _type = "Task" [ 913.241875] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.243030] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebdaf03-ad95-45ff-b632-efa02bad462b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.256765] env[62368]: DEBUG nova.compute.manager [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62368) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 913.257445] env[62368]: DEBUG nova.objects.instance [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.259103] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e722e8-5b8f-c75f-f168-b668c184d851, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.269523] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.270612] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b8c04a9-4ab0-4702-aa84-63cba2bdda6e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.280244] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 913.280244] env[62368]: value = "task-1198765" [ 913.280244] env[62368]: _type = "Task" [ 913.280244] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.283115] env[62368]: DEBUG nova.compute.manager [req-498f8302-afb2-4490-8fe4-c8e290b87707 req-1f413e74-1307-49e0-8090-e142d814665d service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Received event network-vif-plugged-561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.283253] env[62368]: DEBUG oslo_concurrency.lockutils [req-498f8302-afb2-4490-8fe4-c8e290b87707 req-1f413e74-1307-49e0-8090-e142d814665d service nova] Acquiring lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.283411] env[62368]: DEBUG oslo_concurrency.lockutils [req-498f8302-afb2-4490-8fe4-c8e290b87707 req-1f413e74-1307-49e0-8090-e142d814665d service nova] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.283582] env[62368]: DEBUG oslo_concurrency.lockutils [req-498f8302-afb2-4490-8fe4-c8e290b87707 req-1f413e74-1307-49e0-8090-e142d814665d service nova] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.283770] env[62368]: DEBUG nova.compute.manager [req-498f8302-afb2-4490-8fe4-c8e290b87707 req-1f413e74-1307-49e0-8090-e142d814665d service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] No waiting events found dispatching network-vif-plugged-561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 913.283926] env[62368]: WARNING nova.compute.manager [req-498f8302-afb2-4490-8fe4-c8e290b87707 req-1f413e74-1307-49e0-8090-e142d814665d service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Received unexpected event network-vif-plugged-561cb794-4c95-416d-825a-193df8ddc3ca for instance with vm_state building and task_state spawning. [ 913.294401] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198765, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.308975] env[62368]: DEBUG nova.network.neutron [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance_info_cache with network_info: [{"id": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "address": "fa:16:3e:e1:de:4b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be1ff35-0e", "ovs_interfaceid": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.310793] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.370726] env[62368]: DEBUG nova.network.neutron [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Successfully updated port: 561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.390505] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.390755] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2582be8d-e6a2-48fb-bfe8-5667030a6147 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.403654] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 913.403654] env[62368]: value = "task-1198766" [ 913.403654] env[62368]: _type = "Task" [ 913.403654] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.410509] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198766, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.498836] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198763, 'name': CloneVM_Task, 'duration_secs': 1.610862} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.499025] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Created linked-clone VM from snapshot [ 913.499739] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099ad47d-dee2-487d-bf53-36b28205d7f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.508703] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Uploading image bc3d2926-26a5-47b1-b7eb-6a838c226d70 {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 913.542752] env[62368]: DEBUG oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 913.542752] env[62368]: value = "vm-259852" [ 913.542752] env[62368]: _type = "VirtualMachine" [ 913.542752] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 913.543135] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f8056e4d-d2b8-4f00-8b35-ee8593de0851 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.556037] env[62368]: DEBUG oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lease: (returnval){ [ 913.556037] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52aa1715-ea6e-5716-481f-c7e80b83dedf" [ 913.556037] env[62368]: _type = "HttpNfcLease" [ 913.556037] env[62368]: } obtained for exporting VM: (result){ [ 913.556037] env[62368]: value = "vm-259852" [ 913.556037] env[62368]: _type = "VirtualMachine" [ 913.556037] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 913.556426] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the lease: (returnval){ [ 913.556426] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52aa1715-ea6e-5716-481f-c7e80b83dedf" [ 913.556426] env[62368]: _type = "HttpNfcLease" [ 913.556426] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 913.566312] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.566312] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52aa1715-ea6e-5716-481f-c7e80b83dedf" [ 913.566312] env[62368]: _type = "HttpNfcLease" [ 913.566312] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 913.594477] env[62368]: INFO nova.scheduler.client.report [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted allocation for migration a4a55aca-f211-40de-8c6e-9835ad122ae2 [ 913.756243] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e722e8-5b8f-c75f-f168-b668c184d851, 'name': SearchDatastore_Task, 'duration_secs': 0.014676} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.756545] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.756828] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.757086] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.757244] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.757425] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.757695] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf73ad0e-57fa-4bd4-afda-7340bb1cf018 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.762399] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.762586] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be755fae-a93f-4b1d-b2fb-881beaa310f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.770885] env[62368]: DEBUG oslo_vmware.api [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 913.770885] env[62368]: value = "task-1198768" [ 913.770885] env[62368]: _type = "Task" [ 913.770885] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.775795] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.775979] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 913.777047] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8ec667f-44af-42c1-a191-42dfe7096102 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.782949] env[62368]: DEBUG oslo_vmware.api [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198768, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.788035] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 913.788035] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5215cd20-384a-4e6e-514d-f9f6caa6c6af" [ 913.788035] env[62368]: _type = "Task" [ 913.788035] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.793833] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198765, 'name': PowerOffVM_Task, 'duration_secs': 0.30392} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.794993] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 913.795848] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c7ba47-2bf0-4100-ad50-b8585d4cd4f8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.801477] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5215cd20-384a-4e6e-514d-f9f6caa6c6af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.817956] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.822609] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed48459-ce62-42e1-a79d-703711b25959 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.876039] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.876039] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.876039] env[62368]: DEBUG nova.network.neutron [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.913324] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198766, 'name': PowerOffVM_Task, 'duration_secs': 0.148943} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.913324] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 913.913324] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 913.913324] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd22b80-7705-470f-a7d6-017acc08a153 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.919871] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 913.920162] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a6db04e-c1b0-4849-8f23-1773ebd4446b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.951223] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 913.952535] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 913.952535] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Deleting the datastore file [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.952535] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ad314c7-4799-4f7c-b101-9b7f2ae67a34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.960244] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 913.960244] env[62368]: value = "task-1198770" [ 913.960244] env[62368]: _type = "Task" [ 913.960244] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.968946] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198770, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.068834] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 914.068834] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52aa1715-ea6e-5716-481f-c7e80b83dedf" [ 914.068834] env[62368]: _type = "HttpNfcLease" [ 914.068834] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 914.069167] env[62368]: DEBUG oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 914.069167] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52aa1715-ea6e-5716-481f-c7e80b83dedf" [ 914.069167] env[62368]: _type = "HttpNfcLease" [ 914.069167] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 914.069935] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf341bfe-f5e4-4791-9729-353841cda7a5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.088341] env[62368]: DEBUG oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5291ea7f-cc4f-1f99-9fdb-9567ac15f780/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 914.088547] env[62368]: DEBUG oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5291ea7f-cc4f-1f99-9fdb-9567ac15f780/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 914.157616] env[62368]: DEBUG oslo_concurrency.lockutils [None req-eb2dcd34-2f1d-4d0e-8219-c28d4bc5e7bf tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 15.484s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.210205] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b14d23a7-fe00-46c2-bc8f-05c42026c614 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.280316] env[62368]: DEBUG oslo_vmware.api [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198768, 'name': PowerOffVM_Task, 'duration_secs': 0.31899} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.280588] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 914.280767] env[62368]: DEBUG nova.compute.manager [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.281604] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe74230-89d1-4aff-b395-6e7e67648336 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.301812] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5215cd20-384a-4e6e-514d-f9f6caa6c6af, 'name': SearchDatastore_Task, 'duration_secs': 0.020536} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.302955] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4716fc4b-c88e-4880-b700-1c6fd30d60ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.311256] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 914.311256] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52a4cfae-bc13-a26c-c00d-7e596d4c6b7e" [ 914.311256] env[62368]: _type = "Task" [ 914.311256] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.320994] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a4cfae-bc13-a26c-c00d-7e596d4c6b7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.337479] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 914.337479] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f5455ac7-9da0-48e7-862f-8cab8833e9d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.344658] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 914.344658] env[62368]: value = "task-1198771" [ 914.344658] env[62368]: _type = "Task" [ 914.344658] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.349918] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d378e39-79a5-45b6-83bc-bf0718a7a73c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.364628] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198771, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.382805] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf17cea5-e66c-497c-8b56-2dcc2fbeee9d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.390550] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance '5af32858-dc9a-4380-827a-daf384c29bed' progress to 83 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 914.411150] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a72058-9602-4e09-9518-603446f3eabe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.422587] env[62368]: DEBUG nova.network.neutron [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 914.425917] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95159a4f-e899-48bf-ae05-3a99392e6ee3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.461851] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54588633-4763-43c0-a4cb-e19d6d75b258 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.477529] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39ac611-a60d-427f-95cb-b3812195368f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.482378] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198770, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210367} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.485536] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.485740] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 914.485925] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 914.499832] env[62368]: DEBUG nova.compute.provider_tree [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.506498] env[62368]: DEBUG nova.objects.instance [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'flavor' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.725740] env[62368]: DEBUG nova.network.neutron [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating instance_info_cache with network_info: [{"id": "561cb794-4c95-416d-825a-193df8ddc3ca", "address": "fa:16:3e:2b:4e:03", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap561cb794-4c", "ovs_interfaceid": "561cb794-4c95-416d-825a-193df8ddc3ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.796856] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a225a2de-a6ab-4b2e-a70a-5cd0e7bfda79 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.556s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.824721] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a4cfae-bc13-a26c-c00d-7e596d4c6b7e, 'name': SearchDatastore_Task, 'duration_secs': 0.010494} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.825267] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.825708] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8/14c358a0-434d-472a-a382-b5b0e9bf2bd8.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 914.826248] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-223e9b34-c56e-410b-8fbb-94a0f9ddfee5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.836029] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 914.836029] env[62368]: value = "task-1198772" [ 914.836029] env[62368]: _type = "Task" [ 914.836029] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.850258] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198772, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.860961] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198771, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.896897] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 914.897365] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa564715-6349-4f7b-aa22-d55b1d85749b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.906197] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 914.906197] env[62368]: value = "task-1198773" [ 914.906197] env[62368]: _type = "Task" [ 914.906197] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.916934] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198773, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.004494] env[62368]: DEBUG nova.scheduler.client.report [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.022066] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.022705] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.022924] env[62368]: DEBUG nova.network.neutron [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.023334] env[62368]: DEBUG nova.objects.instance [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'info_cache' on Instance uuid 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.230665] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.231026] env[62368]: DEBUG nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Instance network_info: |[{"id": "561cb794-4c95-416d-825a-193df8ddc3ca", "address": "fa:16:3e:2b:4e:03", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap561cb794-4c", "ovs_interfaceid": "561cb794-4c95-416d-825a-193df8ddc3ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.231511] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:4e:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '561cb794-4c95-416d-825a-193df8ddc3ca', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.246719] env[62368]: DEBUG oslo.service.loopingcall [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.246981] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 915.247301] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7795df4b-7f48-418a-a9cf-e5bf8773dad8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.269862] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.269862] env[62368]: value = "task-1198774" [ 915.269862] env[62368]: _type = "Task" [ 915.269862] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.281824] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198774, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.350062] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198772, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507122} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.350132] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8/14c358a0-434d-472a-a382-b5b0e9bf2bd8.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 915.350402] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.358610] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79f82e11-460b-4639-9790-b6d1a05cc047 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.370372] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198771, 'name': CreateSnapshot_Task, 'duration_secs': 0.655375} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.373499] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 915.375225] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 915.375225] env[62368]: value = "task-1198775" [ 915.375225] env[62368]: _type = "Task" [ 915.375225] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.376568] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15cbbbfe-9d05-48ec-8707-bb9e516e6133 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.399085] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198775, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.414857] env[62368]: DEBUG nova.compute.manager [req-a022a8fd-75ca-4638-81b2-d7f2d2b83df5 req-5bb13f74-cc21-49e2-966c-5fcc671965d1 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Received event network-changed-561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.415246] env[62368]: DEBUG nova.compute.manager [req-a022a8fd-75ca-4638-81b2-d7f2d2b83df5 req-5bb13f74-cc21-49e2-966c-5fcc671965d1 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Refreshing instance network info cache due to event network-changed-561cb794-4c95-416d-825a-193df8ddc3ca. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 915.415585] env[62368]: DEBUG oslo_concurrency.lockutils [req-a022a8fd-75ca-4638-81b2-d7f2d2b83df5 req-5bb13f74-cc21-49e2-966c-5fcc671965d1 service nova] Acquiring lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.415796] env[62368]: DEBUG oslo_concurrency.lockutils [req-a022a8fd-75ca-4638-81b2-d7f2d2b83df5 req-5bb13f74-cc21-49e2-966c-5fcc671965d1 service nova] Acquired lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.416024] env[62368]: DEBUG nova.network.neutron [req-a022a8fd-75ca-4638-81b2-d7f2d2b83df5 req-5bb13f74-cc21-49e2-966c-5fcc671965d1 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Refreshing network info cache for port 561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 915.425403] env[62368]: DEBUG oslo_vmware.api [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198773, 'name': PowerOnVM_Task, 'duration_secs': 0.511531} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.425963] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 915.426807] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e1c2e1f5-5a2c-4857-9506-e8a9c0fffe8c tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance '5af32858-dc9a-4380-827a-daf384c29bed' progress to 100 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 915.519373] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.520509] env[62368]: DEBUG nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.523171] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.624s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.525584] env[62368]: DEBUG nova.objects.instance [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lazy-loading 'resources' on Instance uuid 55eecf3d-501c-490d-bddd-0211fd082841 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.530510] env[62368]: DEBUG nova.objects.base [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Object Instance<13765305-2e55-4ee8-9a6f-4ae5ee724367> lazy-loaded attributes: flavor,info_cache {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 915.541849] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.542347] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.543663] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.544166] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.545744] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.545744] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.545744] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.545744] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.545744] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.546741] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.546741] env[62368]: DEBUG nova.virt.hardware [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.547443] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c712819-9e0f-43e7-bc2f-26725d323010 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.560124] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f23e1ac-c0e0-4a7d-b000-a295315a471d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.585305] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.595276] env[62368]: DEBUG oslo.service.loopingcall [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.595276] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 915.595276] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f767842-25e0-4f7f-9083-0c5583c2f7c5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.616554] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.616554] env[62368]: value = "task-1198776" [ 915.616554] env[62368]: _type = "Task" [ 915.616554] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.626238] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198776, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.783749] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198774, 'name': CreateVM_Task, 'duration_secs': 0.399546} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.783970] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 915.784758] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.785617] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.786033] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.786340] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afa4773b-77a4-4339-8a67-b062f30e955c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.792480] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 915.792480] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529b7e67-1dcf-59bb-7d8f-9be9bd718159" [ 915.792480] env[62368]: _type = "Task" [ 915.792480] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.803211] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529b7e67-1dcf-59bb-7d8f-9be9bd718159, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.892453] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198775, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077657} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.892826] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.893814] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47eda9db-0d7e-4b54-952f-a27ca2de4fc1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.918834] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 915.930894] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8/14c358a0-434d-472a-a382-b5b0e9bf2bd8.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.931478] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-04fb5c58-6739-4a92-9296-5f8f36648d8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.935436] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12c24b3a-7bee-4dfd-8e7b-b5fa5ff9a4a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.961514] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 915.961514] env[62368]: value = "task-1198778" [ 915.961514] env[62368]: _type = "Task" [ 915.961514] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.963212] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 915.963212] env[62368]: value = "task-1198777" [ 915.963212] env[62368]: _type = "Task" [ 915.963212] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.983029] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198778, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.984274] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198777, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.028904] env[62368]: DEBUG nova.compute.utils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.031109] env[62368]: DEBUG nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 916.031109] env[62368]: DEBUG nova.network.neutron [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 916.128595] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198776, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.152991] env[62368]: DEBUG nova.objects.instance [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.176547] env[62368]: DEBUG nova.policy [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68c2209271e043b698fd3794955f40bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcecae6c09124e9fa4dcd5d63833b0bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.308400] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529b7e67-1dcf-59bb-7d8f-9be9bd718159, 'name': SearchDatastore_Task, 'duration_secs': 0.02983} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.308400] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.308400] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.308877] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.308877] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.308987] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.309636] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87cae618-ae9d-463e-a040-4af0ad1d416e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.319505] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.319705] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 916.322896] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b56a415-5281-4ecc-8c29-92e95d88289a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.332031] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 916.332031] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529d589d-3d19-f952-58cc-38279ec4f97f" [ 916.332031] env[62368]: _type = "Task" [ 916.332031] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.344104] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529d589d-3d19-f952-58cc-38279ec4f97f, 'name': SearchDatastore_Task, 'duration_secs': 0.010841} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.345090] env[62368]: DEBUG nova.network.neutron [req-a022a8fd-75ca-4638-81b2-d7f2d2b83df5 req-5bb13f74-cc21-49e2-966c-5fcc671965d1 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updated VIF entry in instance network info cache for port 561cb794-4c95-416d-825a-193df8ddc3ca. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 916.345651] env[62368]: DEBUG nova.network.neutron [req-a022a8fd-75ca-4638-81b2-d7f2d2b83df5 req-5bb13f74-cc21-49e2-966c-5fcc671965d1 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating instance_info_cache with network_info: [{"id": "561cb794-4c95-416d-825a-193df8ddc3ca", "address": "fa:16:3e:2b:4e:03", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap561cb794-4c", "ovs_interfaceid": "561cb794-4c95-416d-825a-193df8ddc3ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.347981] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ec8bde5-8d85-41fe-a0a0-a02bfec586fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.358663] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 916.358663] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52028839-bbdb-3276-42b5-312215fe5fa4" [ 916.358663] env[62368]: _type = "Task" [ 916.358663] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.369571] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52028839-bbdb-3276-42b5-312215fe5fa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.374901] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9294ec-4330-411a-acd1-0b5d4c1e0670 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.384487] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f21d96e-7e8c-4e93-9f41-abb2623bc005 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.420533] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2de65b-2efd-40f4-88f8-0a6e8a1ff164 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.429316] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2661d8-ee89-4b3a-b244-1eb797792939 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.445922] env[62368]: DEBUG nova.compute.provider_tree [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.476866] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198778, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.483337] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198777, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.515656] env[62368]: DEBUG nova.network.neutron [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [{"id": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "address": "fa:16:3e:5e:9e:c0", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11ce4c60-0d", "ovs_interfaceid": "11ce4c60-0df0-4c43-bbe4-6cc53717309b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.534491] env[62368]: DEBUG nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.628469] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198776, 'name': CreateVM_Task, 'duration_secs': 0.592962} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.628651] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 916.632025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.632025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.632025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.632025] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1590a8ac-4af9-4bef-8ead-5c1db92c3188 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.638372] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 916.638372] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52618d02-fb44-e15e-316b-f828fc042bb8" [ 916.638372] env[62368]: _type = "Task" [ 916.638372] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.646675] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52618d02-fb44-e15e-316b-f828fc042bb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.658417] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.658630] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.658824] env[62368]: DEBUG nova.network.neutron [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 916.659036] env[62368]: DEBUG nova.objects.instance [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'info_cache' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.718975] env[62368]: DEBUG nova.network.neutron [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Successfully created port: 14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.849396] env[62368]: DEBUG oslo_concurrency.lockutils [req-a022a8fd-75ca-4638-81b2-d7f2d2b83df5 req-5bb13f74-cc21-49e2-966c-5fcc671965d1 service nova] Releasing lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.872347] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52028839-bbdb-3276-42b5-312215fe5fa4, 'name': SearchDatastore_Task, 'duration_secs': 0.011164} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.872347] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.872347] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6/3e09be20-ce79-4bfa-bf84-2877f4e534e6.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 916.875076] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-286d1f2d-ddde-4f2f-8ca4-43eb70f49882 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.884255] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 916.884255] env[62368]: value = "task-1198779" [ 916.884255] env[62368]: _type = "Task" [ 916.884255] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.893915] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198779, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.950058] env[62368]: DEBUG nova.scheduler.client.report [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.985872] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198777, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.989611] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198778, 'name': ReconfigVM_Task, 'duration_secs': 0.606849} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.989961] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8/14c358a0-434d-472a-a382-b5b0e9bf2bd8.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.995314] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed7e77e0-2d95-4335-bad9-68259d710a17 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.004274] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 917.004274] env[62368]: value = "task-1198780" [ 917.004274] env[62368]: _type = "Task" [ 917.004274] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.020753] env[62368]: DEBUG oslo_concurrency.lockutils [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-13765305-2e55-4ee8-9a6f-4ae5ee724367" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.022484] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198780, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.152188] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52618d02-fb44-e15e-316b-f828fc042bb8, 'name': SearchDatastore_Task, 'duration_secs': 0.010959} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.152642] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.153680] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.154042] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.154308] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.154549] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.154970] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74d3a610-977c-4817-990c-56c8da2ffce5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.162161] env[62368]: DEBUG nova.objects.base [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Object Instance<58157ab1-80a4-427c-812b-f6fde1f8db68> lazy-loaded attributes: flavor,info_cache {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 917.180333] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.180712] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 917.181806] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fd08bbe-2375-4d9e-b0da-fd5d898b22dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.193392] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 917.193392] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52593bd5-761d-1e6a-9c09-0811712d50df" [ 917.193392] env[62368]: _type = "Task" [ 917.193392] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.205457] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52593bd5-761d-1e6a-9c09-0811712d50df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.399315] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198779, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5101} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.402286] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6/3e09be20-ce79-4bfa-bf84-2877f4e534e6.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 917.402286] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.402286] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-943ea7e9-4c7a-48ad-9e24-5236104d6011 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.412273] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 917.412273] env[62368]: value = "task-1198781" [ 917.412273] env[62368]: _type = "Task" [ 917.412273] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.433269] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198781, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.456270] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.458779] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.148s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.459039] env[62368]: DEBUG nova.objects.instance [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lazy-loading 'resources' on Instance uuid ad21bf3a-068e-44a7-b58f-2978a631b972 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.482769] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198777, 'name': CloneVM_Task, 'duration_secs': 1.386076} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.484013] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Created linked-clone VM from snapshot [ 917.487766] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b391bc9-08e8-4726-82e0-90ba0304e312 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.491172] env[62368]: INFO nova.scheduler.client.report [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Deleted allocations for instance 55eecf3d-501c-490d-bddd-0211fd082841 [ 917.500363] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Uploading image c4f6a28c-a893-4200-ba2f-b1e0d47d22df {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 917.520286] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198780, 'name': Rename_Task, 'duration_secs': 0.207622} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.522805] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 917.524251] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 917.524485] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bcd9997-a1b1-4b85-8097-dbc0c03ce473 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.526773] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11ca06c3-404b-41ae-8609-198cb88b568c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.531063] env[62368]: DEBUG oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 917.531063] env[62368]: value = "vm-259857" [ 917.531063] env[62368]: _type = "VirtualMachine" [ 917.531063] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 917.531752] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b76d16be-28ea-49a0-83fa-8e69242e8f3f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.536240] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 917.536240] env[62368]: value = "task-1198782" [ 917.536240] env[62368]: _type = "Task" [ 917.536240] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.541487] env[62368]: DEBUG oslo_vmware.api [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 917.541487] env[62368]: value = "task-1198783" [ 917.541487] env[62368]: _type = "Task" [ 917.541487] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.546014] env[62368]: DEBUG nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.555736] env[62368]: DEBUG oslo_vmware.api [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198783, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.561562] env[62368]: DEBUG oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lease: (returnval){ [ 917.561562] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522477ef-e6f9-672d-901d-39b2c1f66097" [ 917.561562] env[62368]: _type = "HttpNfcLease" [ 917.561562] env[62368]: } obtained for exporting VM: (result){ [ 917.561562] env[62368]: value = "vm-259857" [ 917.561562] env[62368]: _type = "VirtualMachine" [ 917.561562] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 917.561562] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the lease: (returnval){ [ 917.561562] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522477ef-e6f9-672d-901d-39b2c1f66097" [ 917.561562] env[62368]: _type = "HttpNfcLease" [ 917.561562] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 917.561562] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198782, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.568871] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 917.568871] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522477ef-e6f9-672d-901d-39b2c1f66097" [ 917.568871] env[62368]: _type = "HttpNfcLease" [ 917.568871] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 917.571026] env[62368]: DEBUG oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 917.571026] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522477ef-e6f9-672d-901d-39b2c1f66097" [ 917.571026] env[62368]: _type = "HttpNfcLease" [ 917.571026] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 917.572346] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2758db64-3b78-48b4-9972-2e44534d62da {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.581466] env[62368]: DEBUG oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5201ba4f-ea86-999f-6868-89acad5b0a80/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 917.581664] env[62368]: DEBUG oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5201ba4f-ea86-999f-6868-89acad5b0a80/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 917.653640] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.654147] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.654147] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.654581] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.654790] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.654949] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.655213] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.655419] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.655789] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.655995] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.656196] env[62368]: DEBUG nova.virt.hardware [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.657124] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea507802-b939-415d-9c87-befb34a2a948 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.668715] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51f5684-a64f-4518-aed4-7acf496827ef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.702018] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-89ff9a86-7ea8-46ba-ad91-09a70da70837 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.711384] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52593bd5-761d-1e6a-9c09-0811712d50df, 'name': SearchDatastore_Task, 'duration_secs': 0.050242} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.711384] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3767e60-50e5-4d07-b292-e344647ba20b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.719599] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 917.719599] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e539ff-0da6-f03a-b7dc-15a9d340056c" [ 917.719599] env[62368]: _type = "Task" [ 917.719599] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.728555] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e539ff-0da6-f03a-b7dc-15a9d340056c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.923407] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "5af32858-dc9a-4380-827a-daf384c29bed" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.923594] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.923948] env[62368]: DEBUG nova.compute.manager [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Going to confirm migration 3 {{(pid=62368) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 917.925322] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198781, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085856} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.925589] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.926739] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8934a84-9aa4-4493-840a-338183437c64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.954990] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6/3e09be20-ce79-4bfa-bf84-2877f4e534e6.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.959111] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdd5b809-6ed5-4d61-820e-4d985cb53c0c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.005836] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 918.005836] env[62368]: value = "task-1198785" [ 918.005836] env[62368]: _type = "Task" [ 918.005836] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.014251] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2b372555-40a2-485e-97d7-3eb41a3ee651 tempest-ImagesTestJSON-763197622 tempest-ImagesTestJSON-763197622-project-member] Lock "55eecf3d-501c-490d-bddd-0211fd082841" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.352s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.023712] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198785, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.056836] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198782, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.064879] env[62368]: DEBUG oslo_vmware.api [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198783, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.077767] env[62368]: DEBUG nova.network.neutron [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updating instance_info_cache with network_info: [{"id": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "address": "fa:16:3e:23:0d:cc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46fba643-18", "ovs_interfaceid": "46fba643-185a-4fb2-ab24-dd4350aa5ac4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.235495] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e539ff-0da6-f03a-b7dc-15a9d340056c, 'name': SearchDatastore_Task, 'duration_secs': 0.013136} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.240345] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.241137] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc/7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 918.242359] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50e158c5-2ba6-4aee-9903-8e20b3f5b2f4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.253147] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 918.253147] env[62368]: value = "task-1198786" [ 918.253147] env[62368]: _type = "Task" [ 918.253147] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.269938] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198786, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.349852] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216e830b-8fc3-4e86-bb3f-691455e1fd0f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.361857] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279e86df-e1fd-4a43-8983-435c94246a2a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.402121] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f87c077-b2d8-4243-9cae-4f89e7802038 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.413069] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8455ba5-7e8b-4ad8-ad4c-bd60f3598198 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.421186] env[62368]: DEBUG nova.compute.manager [req-860a72a7-4fde-4795-bd62-c44b45c813b1 req-48ee8780-5c75-4b9d-8190-b06b7ce2eec5 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received event network-vif-plugged-14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.421582] env[62368]: DEBUG oslo_concurrency.lockutils [req-860a72a7-4fde-4795-bd62-c44b45c813b1 req-48ee8780-5c75-4b9d-8190-b06b7ce2eec5 service nova] Acquiring lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.421878] env[62368]: DEBUG oslo_concurrency.lockutils [req-860a72a7-4fde-4795-bd62-c44b45c813b1 req-48ee8780-5c75-4b9d-8190-b06b7ce2eec5 service nova] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.422090] env[62368]: DEBUG oslo_concurrency.lockutils [req-860a72a7-4fde-4795-bd62-c44b45c813b1 req-48ee8780-5c75-4b9d-8190-b06b7ce2eec5 service nova] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.422332] env[62368]: DEBUG nova.compute.manager [req-860a72a7-4fde-4795-bd62-c44b45c813b1 req-48ee8780-5c75-4b9d-8190-b06b7ce2eec5 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] No waiting events found dispatching network-vif-plugged-14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.422593] env[62368]: WARNING nova.compute.manager [req-860a72a7-4fde-4795-bd62-c44b45c813b1 req-48ee8780-5c75-4b9d-8190-b06b7ce2eec5 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received unexpected event network-vif-plugged-14840bf2-7da7-4d91-aa96-02a1771fb7d1 for instance with vm_state building and task_state spawning. [ 918.435824] env[62368]: DEBUG nova.compute.provider_tree [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.522314] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198785, 'name': ReconfigVM_Task, 'duration_secs': 0.407568} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.522748] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6/3e09be20-ce79-4bfa-bf84-2877f4e534e6.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.523860] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c49f941-c656-4d2e-8c85-e3574395e24d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.532584] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 918.532584] env[62368]: value = "task-1198787" [ 918.532584] env[62368]: _type = "Task" [ 918.532584] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.549069] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198787, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.557982] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198782, 'name': PowerOnVM_Task, 'duration_secs': 0.804861} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.558325] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.558413] env[62368]: DEBUG nova.compute.manager [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.559930] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2f9dc6-de78-4e3c-85fb-b6cd18e28afc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.567740] env[62368]: DEBUG oslo_vmware.api [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198783, 'name': PowerOnVM_Task, 'duration_secs': 0.547875} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.571182] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.571869] env[62368]: DEBUG nova.compute.manager [None req-0874169b-cc32-4cad-8c37-e335c45062e1 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.575897] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f3c028-2dcb-4adb-bea2-f5de9fef62a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.585842] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Releasing lock "refresh_cache-58157ab1-80a4-427c-812b-f6fde1f8db68" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.629909] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.630307] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.630658] env[62368]: DEBUG nova.network.neutron [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 918.631038] env[62368]: DEBUG nova.objects.instance [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lazy-loading 'info_cache' on Instance uuid 5af32858-dc9a-4380-827a-daf384c29bed {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.767551] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198786, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.840754] env[62368]: DEBUG nova.network.neutron [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Successfully updated port: 14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.941615] env[62368]: DEBUG nova.scheduler.client.report [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.043927] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198787, 'name': Rename_Task, 'duration_secs': 0.36882} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.044346] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 919.044642] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50d9c154-3ae9-4432-92ef-e82dc54f28b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.052906] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 919.052906] env[62368]: value = "task-1198788" [ 919.052906] env[62368]: _type = "Task" [ 919.052906] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.068533] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198788, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.089806] env[62368]: INFO nova.compute.manager [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] bringing vm to original state: 'stopped' [ 919.097034] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 919.097034] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-659bf9c0-43ee-4f0e-aea2-6d8b77ce410a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.110612] env[62368]: DEBUG oslo_vmware.api [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 919.110612] env[62368]: value = "task-1198789" [ 919.110612] env[62368]: _type = "Task" [ 919.110612] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.120065] env[62368]: DEBUG oslo_vmware.api [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198789, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.269660] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198786, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541934} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.270674] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc/7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 919.270674] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.270893] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f1736c97-359d-4a8d-b26a-08e24195d52f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.280530] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 919.280530] env[62368]: value = "task-1198790" [ 919.280530] env[62368]: _type = "Task" [ 919.280530] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.292995] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198790, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.344784] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.344950] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.345045] env[62368]: DEBUG nova.network.neutron [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 919.448586] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.989s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.479513] env[62368]: INFO nova.scheduler.client.report [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted allocations for instance ad21bf3a-068e-44a7-b58f-2978a631b972 [ 919.563746] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.564030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.564270] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.564693] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.564693] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.566883] env[62368]: INFO nova.compute.manager [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Terminating instance [ 919.570199] env[62368]: DEBUG nova.compute.manager [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.570199] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 919.574812] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4b7120-bb43-4102-a6c0-d314e322057b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.577676] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198788, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.583553] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 919.583821] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8809786d-9e95-4692-bf19-1896a8cb2ef9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.592560] env[62368]: DEBUG oslo_vmware.api [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 919.592560] env[62368]: value = "task-1198791" [ 919.592560] env[62368]: _type = "Task" [ 919.592560] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.606556] env[62368]: DEBUG oslo_vmware.api [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198791, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.625324] env[62368]: DEBUG oslo_vmware.api [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198789, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.793542] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198790, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117407} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.793958] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.794989] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22193896-77b4-41ef-baed-adbf39e1cf93 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.818203] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc/7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.821696] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-671e6a6f-7e0c-458b-ad7c-0345642f39e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.850080] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 919.850080] env[62368]: value = "task-1198792" [ 919.850080] env[62368]: _type = "Task" [ 919.850080] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.865401] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198792, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.918057] env[62368]: DEBUG nova.network.neutron [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.961880] env[62368]: DEBUG nova.network.neutron [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance_info_cache with network_info: [{"id": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "address": "fa:16:3e:e1:de:4b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be1ff35-0e", "ovs_interfaceid": "8be1ff35-0e69-42f1-a582-16bf1f496a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.990689] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5b0ba98b-677f-41ae-a83b-122a9227bcb7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "ad21bf3a-068e-44a7-b58f-2978a631b972" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.721s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.064264] env[62368]: DEBUG oslo_vmware.api [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198788, 'name': PowerOnVM_Task, 'duration_secs': 0.678793} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.066512] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 920.066512] env[62368]: INFO nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Took 7.45 seconds to spawn the instance on the hypervisor. [ 920.066512] env[62368]: DEBUG nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.066512] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3655459e-76d0-46a0-91d2-87b02be64df8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.097531] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.097930] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.098233] env[62368]: DEBUG nova.compute.manager [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.099204] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6acf95ad-3488-40d7-aadd-b1554ddce15b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.110182] env[62368]: DEBUG oslo_vmware.api [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198791, 'name': PowerOffVM_Task, 'duration_secs': 0.2731} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.112223] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 920.112423] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 920.112721] env[62368]: DEBUG nova.compute.manager [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62368) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 920.118060] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8741f32b-a030-468b-a0fc-48e23d5b34df {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.123602] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 920.123869] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d7afe6b-51df-4c96-9ede-2dcf8c5bb09a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.132283] env[62368]: DEBUG oslo_vmware.api [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198789, 'name': PowerOnVM_Task, 'duration_secs': 0.796935} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.133838] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 920.134236] env[62368]: DEBUG nova.compute.manager [None req-1798e0be-cdba-41e9-b499-13f6dd0ba27c tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.134597] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 920.134597] env[62368]: value = "task-1198794" [ 920.134597] env[62368]: _type = "Task" [ 920.134597] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.135616] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e253b0b-62e2-4cc9-8694-9a3948efd741 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.151927] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198794, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.173015] env[62368]: DEBUG nova.network.neutron [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.216514] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 920.216938] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 920.217283] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleting the datastore file [datastore2] 13765305-2e55-4ee8-9a6f-4ae5ee724367 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.217644] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb959996-0bd9-4208-bb76-6532d011663b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.226035] env[62368]: DEBUG oslo_vmware.api [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 920.226035] env[62368]: value = "task-1198795" [ 920.226035] env[62368]: _type = "Task" [ 920.226035] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.235612] env[62368]: DEBUG oslo_vmware.api [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198795, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.366366] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198792, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.469679] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-5af32858-dc9a-4380-827a-daf384c29bed" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.470789] env[62368]: DEBUG nova.objects.instance [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lazy-loading 'migration_context' on Instance uuid 5af32858-dc9a-4380-827a-daf384c29bed {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.576221] env[62368]: DEBUG nova.compute.manager [req-5f94dbf0-9aa5-4b01-b60d-bd9474d71d4e req-624c90d3-6f32-4c9f-9e00-a91e9c89c7f0 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received event network-changed-14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.576221] env[62368]: DEBUG nova.compute.manager [req-5f94dbf0-9aa5-4b01-b60d-bd9474d71d4e req-624c90d3-6f32-4c9f-9e00-a91e9c89c7f0 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing instance network info cache due to event network-changed-14840bf2-7da7-4d91-aa96-02a1771fb7d1. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 920.576221] env[62368]: DEBUG oslo_concurrency.lockutils [req-5f94dbf0-9aa5-4b01-b60d-bd9474d71d4e req-624c90d3-6f32-4c9f-9e00-a91e9c89c7f0 service nova] Acquiring lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.590102] env[62368]: INFO nova.compute.manager [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Took 18.67 seconds to build instance. [ 920.651186] env[62368]: DEBUG oslo_vmware.api [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198794, 'name': PowerOffVM_Task, 'duration_secs': 0.24351} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.651186] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 920.651186] env[62368]: DEBUG nova.compute.manager [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.652382] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3617c1-1fc6-47cd-9c50-0d8114d7d1db {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.674197] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.677022] env[62368]: DEBUG nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Instance network_info: |[{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.677022] env[62368]: DEBUG oslo_concurrency.lockutils [req-5f94dbf0-9aa5-4b01-b60d-bd9474d71d4e req-624c90d3-6f32-4c9f-9e00-a91e9c89c7f0 service nova] Acquired lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.677022] env[62368]: DEBUG nova.network.neutron [req-5f94dbf0-9aa5-4b01-b60d-bd9474d71d4e req-624c90d3-6f32-4c9f-9e00-a91e9c89c7f0 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing network info cache for port 14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 920.677022] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:85:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14840bf2-7da7-4d91-aa96-02a1771fb7d1', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.687105] env[62368]: DEBUG oslo.service.loopingcall [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.688534] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 920.691060] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91777918-8feb-4847-917c-861fb49bc3e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.713273] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.713273] env[62368]: value = "task-1198796" [ 920.713273] env[62368]: _type = "Task" [ 920.713273] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.724018] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198796, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.737547] env[62368]: DEBUG oslo_vmware.api [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198795, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.404719} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.737922] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.738182] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 920.738394] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 920.738637] env[62368]: INFO nova.compute.manager [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Took 1.17 seconds to destroy the instance on the hypervisor. [ 920.738926] env[62368]: DEBUG oslo.service.loopingcall [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.739211] env[62368]: DEBUG nova.compute.manager [-] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 920.739500] env[62368]: DEBUG nova.network.neutron [-] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 920.874751] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198792, 'name': ReconfigVM_Task, 'duration_secs': 0.581402} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.875387] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc/7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.876383] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a29b3e1a-b2b0-479f-9227-164dd737700c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.886129] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 920.886129] env[62368]: value = "task-1198797" [ 920.886129] env[62368]: _type = "Task" [ 920.886129] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.896461] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198797, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.974488] env[62368]: DEBUG nova.objects.base [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Object Instance<5af32858-dc9a-4380-827a-daf384c29bed> lazy-loaded attributes: info_cache,migration_context {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 920.975552] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2d1d4e-d304-4196-8176-72092cd40c0b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.996614] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb3f754e-e000-4082-8dc3-93ff9f1d235b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.003496] env[62368]: DEBUG oslo_vmware.api [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 921.003496] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524191dd-32d2-2799-dfdf-d049df108af9" [ 921.003496] env[62368]: _type = "Task" [ 921.003496] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.012662] env[62368]: DEBUG oslo_vmware.api [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524191dd-32d2-2799-dfdf-d049df108af9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.090845] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27295dbc-8099-4361-b564-1ea9d7d45cd4 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.175s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.172061] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.074s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.231802] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198796, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.398792] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198797, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.455897] env[62368]: DEBUG nova.network.neutron [req-5f94dbf0-9aa5-4b01-b60d-bd9474d71d4e req-624c90d3-6f32-4c9f-9e00-a91e9c89c7f0 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updated VIF entry in instance network info cache for port 14840bf2-7da7-4d91-aa96-02a1771fb7d1. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 921.457368] env[62368]: DEBUG nova.network.neutron [req-5f94dbf0-9aa5-4b01-b60d-bd9474d71d4e req-624c90d3-6f32-4c9f-9e00-a91e9c89c7f0 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.514678] env[62368]: DEBUG oslo_vmware.api [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524191dd-32d2-2799-dfdf-d049df108af9, 'name': SearchDatastore_Task, 'duration_secs': 0.016848} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.515019] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.515281] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.616305] env[62368]: DEBUG nova.compute.manager [req-7b7a5b03-98c9-45d0-b1c7-d5326b354689 req-3e6fb3c2-3b2e-46a1-bab8-a828e8bd892b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Received event network-vif-deleted-11ce4c60-0df0-4c43-bbe4-6cc53717309b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.616471] env[62368]: INFO nova.compute.manager [req-7b7a5b03-98c9-45d0-b1c7-d5326b354689 req-3e6fb3c2-3b2e-46a1-bab8-a828e8bd892b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Neutron deleted interface 11ce4c60-0df0-4c43-bbe4-6cc53717309b; detaching it from the instance and deleting it from the info cache [ 921.616718] env[62368]: DEBUG nova.network.neutron [req-7b7a5b03-98c9-45d0-b1c7-d5326b354689 req-3e6fb3c2-3b2e-46a1-bab8-a828e8bd892b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.689090] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.729810] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198796, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.898934] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198797, 'name': Rename_Task, 'duration_secs': 0.894391} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.898934] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 921.899226] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68ec50c4-7d17-4adc-877a-2f82ee302a65 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.906770] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 921.906770] env[62368]: value = "task-1198798" [ 921.906770] env[62368]: _type = "Task" [ 921.906770] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.915753] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.963755] env[62368]: DEBUG oslo_concurrency.lockutils [req-5f94dbf0-9aa5-4b01-b60d-bd9474d71d4e req-624c90d3-6f32-4c9f-9e00-a91e9c89c7f0 service nova] Releasing lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.964245] env[62368]: DEBUG nova.network.neutron [-] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.123361] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7540849-2a58-4a89-81cd-25bb7a8140d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.133728] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ac88d0-abd5-41a5-b431-4a59201d39c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.189945] env[62368]: DEBUG nova.compute.manager [req-7b7a5b03-98c9-45d0-b1c7-d5326b354689 req-3e6fb3c2-3b2e-46a1-bab8-a828e8bd892b service nova] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Detach interface failed, port_id=11ce4c60-0df0-4c43-bbe4-6cc53717309b, reason: Instance 13765305-2e55-4ee8-9a6f-4ae5ee724367 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 922.229749] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198796, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.280856] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd44afe2-f098-4e56-9ea8-213a06845fda {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.289894] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a455efe-39b8-426f-8396-079bc3c69a20 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.328697] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99feb297-040e-4ad7-8124-b19925c4e938 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.338466] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6aa967-6369-4f31-a452-4d4f5e6316c9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.353602] env[62368]: DEBUG nova.compute.provider_tree [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.418120] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198798, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.467625] env[62368]: INFO nova.compute.manager [-] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Took 1.73 seconds to deallocate network for instance. [ 922.727807] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198796, 'name': CreateVM_Task, 'duration_secs': 1.610393} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.728207] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 922.728657] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.728832] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.729236] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.729448] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abe9f84f-0c4c-47d1-8263-77dc9920f12f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.735289] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 922.735289] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fd0a04-59fb-c562-2f5c-d4bb4c20bd5d" [ 922.735289] env[62368]: _type = "Task" [ 922.735289] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.744960] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fd0a04-59fb-c562-2f5c-d4bb4c20bd5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.857480] env[62368]: DEBUG nova.scheduler.client.report [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.921219] env[62368]: DEBUG oslo_vmware.api [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198798, 'name': PowerOnVM_Task, 'duration_secs': 0.557255} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.921772] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 922.922046] env[62368]: DEBUG nova.compute.manager [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.922952] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d34a80-5671-4983-87a8-69b43facb8ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.976023] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.247309] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fd0a04-59fb-c562-2f5c-d4bb4c20bd5d, 'name': SearchDatastore_Task, 'duration_secs': 0.01742} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.247644] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.247924] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.248255] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.248344] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.248562] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.248863] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59f20b1c-2796-4eb8-86c8-45390d2d6ccf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.259098] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.259346] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 923.260235] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97f88519-366d-4801-b3d4-94cb71a753e1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.266483] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 923.266483] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5202b66c-462e-a72b-9a74-02d8aff2df1e" [ 923.266483] env[62368]: _type = "Task" [ 923.266483] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.275774] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5202b66c-462e-a72b-9a74-02d8aff2df1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.442588] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.777503] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5202b66c-462e-a72b-9a74-02d8aff2df1e, 'name': SearchDatastore_Task, 'duration_secs': 0.017626} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.778362] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4d02f89-1b23-4775-84b8-71434fc056c1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.785027] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 923.785027] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f09430-b234-8ce6-70e9-78477a3fa294" [ 923.785027] env[62368]: _type = "Task" [ 923.785027] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.793772] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f09430-b234-8ce6-70e9-78477a3fa294, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.868850] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.353s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.871980] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.183s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.872118] env[62368]: DEBUG nova.objects.instance [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 924.244167] env[62368]: DEBUG oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5291ea7f-cc4f-1f99-9fdb-9567ac15f780/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 924.245389] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ef0993-e643-4262-8d21-3ca3da7838bc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.253598] env[62368]: DEBUG oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5291ea7f-cc4f-1f99-9fdb-9567ac15f780/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 924.253893] env[62368]: ERROR oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5291ea7f-cc4f-1f99-9fdb-9567ac15f780/disk-0.vmdk due to incomplete transfer. [ 924.254046] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fbb22b43-d6c9-4de2-ba48-0d2f18367ea5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.264255] env[62368]: DEBUG oslo_vmware.rw_handles [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5291ea7f-cc4f-1f99-9fdb-9567ac15f780/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 924.264255] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Uploaded image bc3d2926-26a5-47b1-b7eb-6a838c226d70 to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 924.266834] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 924.267185] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fd6ed8d8-09be-4257-b018-aacf706cb886 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.275912] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 924.275912] env[62368]: value = "task-1198799" [ 924.275912] env[62368]: _type = "Task" [ 924.275912] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.287425] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198799, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.297317] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f09430-b234-8ce6-70e9-78477a3fa294, 'name': SearchDatastore_Task, 'duration_secs': 0.01431} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.297648] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.297908] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 5633292c-9a74-4c2d-893b-6823f4a587a2/5633292c-9a74-4c2d-893b-6823f4a587a2.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 924.298223] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e376bf6-5b2e-41d5-935e-0b396aa438cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.307061] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 924.307061] env[62368]: value = "task-1198800" [ 924.307061] env[62368]: _type = "Task" [ 924.307061] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.317236] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198800, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.445077] env[62368]: INFO nova.scheduler.client.report [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted allocation for migration db16e018-9b81-40e4-8b6b-0ac690c9f100 [ 924.784650] env[62368]: DEBUG nova.compute.manager [req-e38f6634-93c5-42c6-840b-9511652efd7f req-4f424492-1825-4fe4-adf0-bff769217ab0 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Received event network-changed-561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.784946] env[62368]: DEBUG nova.compute.manager [req-e38f6634-93c5-42c6-840b-9511652efd7f req-4f424492-1825-4fe4-adf0-bff769217ab0 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Refreshing instance network info cache due to event network-changed-561cb794-4c95-416d-825a-193df8ddc3ca. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 924.785177] env[62368]: DEBUG oslo_concurrency.lockutils [req-e38f6634-93c5-42c6-840b-9511652efd7f req-4f424492-1825-4fe4-adf0-bff769217ab0 service nova] Acquiring lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.785400] env[62368]: DEBUG oslo_concurrency.lockutils [req-e38f6634-93c5-42c6-840b-9511652efd7f req-4f424492-1825-4fe4-adf0-bff769217ab0 service nova] Acquired lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.785632] env[62368]: DEBUG nova.network.neutron [req-e38f6634-93c5-42c6-840b-9511652efd7f req-4f424492-1825-4fe4-adf0-bff769217ab0 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Refreshing network info cache for port 561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 924.796799] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198799, 'name': Destroy_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.820861] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198800, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.886367] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a75c948-7e58-4961-b0dd-3d838581d45d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.887745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.912s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.887960] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.890840] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.448s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.890840] env[62368]: DEBUG nova.objects.instance [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 924.921203] env[62368]: INFO nova.scheduler.client.report [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted allocations for instance 13765305-2e55-4ee8-9a6f-4ae5ee724367 [ 924.954176] env[62368]: DEBUG oslo_concurrency.lockutils [None req-45563cd7-20fd-4b07-9468-e24c6f4d58f6 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.030s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.993423] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Acquiring lock "4785157b-da21-4ed9-aae2-b071b42688e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.993812] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Lock "4785157b-da21-4ed9-aae2-b071b42688e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.169132] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "e427d283-1f8f-4c7a-869c-4cd48a595470" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.169413] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "e427d283-1f8f-4c7a-869c-4cd48a595470" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.287655] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198799, 'name': Destroy_Task, 'duration_secs': 0.610511} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.288479] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Destroyed the VM [ 925.288733] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 925.289052] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-459cd8ec-5d7a-4372-b75b-8cc135b76a22 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.299827] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 925.299827] env[62368]: value = "task-1198804" [ 925.299827] env[62368]: _type = "Task" [ 925.299827] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.308669] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198804, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.320751] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198800, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648293} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.320751] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 5633292c-9a74-4c2d-893b-6823f4a587a2/5633292c-9a74-4c2d-893b-6823f4a587a2.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.321082] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.321223] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3897fce5-666a-442c-80df-c68f97a43dcf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.329706] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 925.329706] env[62368]: value = "task-1198805" [ 925.329706] env[62368]: _type = "Task" [ 925.329706] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.341013] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198805, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.432082] env[62368]: DEBUG oslo_concurrency.lockutils [None req-27606fc1-be0c-4153-a898-9cd819bcef2c tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "13765305-2e55-4ee8-9a6f-4ae5ee724367" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.868s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.496563] env[62368]: DEBUG nova.compute.manager [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 925.634502] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.634786] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.635019] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.635322] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.635527] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.639811] env[62368]: DEBUG nova.network.neutron [req-e38f6634-93c5-42c6-840b-9511652efd7f req-4f424492-1825-4fe4-adf0-bff769217ab0 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updated VIF entry in instance network info cache for port 561cb794-4c95-416d-825a-193df8ddc3ca. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.640212] env[62368]: DEBUG nova.network.neutron [req-e38f6634-93c5-42c6-840b-9511652efd7f req-4f424492-1825-4fe4-adf0-bff769217ab0 service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating instance_info_cache with network_info: [{"id": "561cb794-4c95-416d-825a-193df8ddc3ca", "address": "fa:16:3e:2b:4e:03", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap561cb794-4c", "ovs_interfaceid": "561cb794-4c95-416d-825a-193df8ddc3ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.643593] env[62368]: INFO nova.compute.manager [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Terminating instance [ 925.646061] env[62368]: DEBUG nova.compute.manager [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.646061] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 925.646061] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91c63e7-f00d-46ec-98bf-d19d8864eabf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.655687] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 925.656768] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a66d211-703e-4913-a1dd-483985bfa5cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.672278] env[62368]: DEBUG nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 925.739583] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 925.739948] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 925.741990] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleting the datastore file [datastore2] 14c358a0-434d-472a-a382-b5b0e9bf2bd8 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.742395] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50a874fc-4902-4d82-baf8-fe329806188e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.752170] env[62368]: DEBUG oslo_vmware.api [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 925.752170] env[62368]: value = "task-1198807" [ 925.752170] env[62368]: _type = "Task" [ 925.752170] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.763646] env[62368]: DEBUG oslo_vmware.api [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198807, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.777185] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.777454] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.777699] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.777901] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.778092] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.780872] env[62368]: INFO nova.compute.manager [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Terminating instance [ 925.782662] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "refresh_cache-7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.782863] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquired lock "refresh_cache-7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.783454] env[62368]: DEBUG nova.network.neutron [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 925.811463] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198804, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.841492] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198805, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083795} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.841758] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.842570] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f0295a-8084-4ea5-9204-00892377d778 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.870383] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 5633292c-9a74-4c2d-893b-6823f4a587a2/5633292c-9a74-4c2d-893b-6823f4a587a2.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.871639] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4627af6-29ff-4e58-b64e-986494a3e566 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.895998] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 925.895998] env[62368]: value = "task-1198808" [ 925.895998] env[62368]: _type = "Task" [ 925.895998] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.905535] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d5c81f13-e040-4f09-8c14-99a74a0a92ac tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.906797] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198808, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.014172] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.014172] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.015217] env[62368]: INFO nova.compute.claims [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.145022] env[62368]: DEBUG oslo_concurrency.lockutils [req-e38f6634-93c5-42c6-840b-9511652efd7f req-4f424492-1825-4fe4-adf0-bff769217ab0 service nova] Releasing lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.197949] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.264562] env[62368]: DEBUG oslo_vmware.api [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198807, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.391107} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.265278] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.265278] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 926.265278] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 926.265473] env[62368]: INFO nova.compute.manager [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Took 0.62 seconds to destroy the instance on the hypervisor. [ 926.265862] env[62368]: DEBUG oslo.service.loopingcall [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.266098] env[62368]: DEBUG nova.compute.manager [-] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.266195] env[62368]: DEBUG nova.network.neutron [-] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 926.306333] env[62368]: DEBUG nova.network.neutron [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 926.312935] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198804, 'name': RemoveSnapshot_Task, 'duration_secs': 0.838536} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.313241] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 926.313517] env[62368]: DEBUG nova.compute.manager [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.314636] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c84621-6d9b-49c9-b5b0-928d46cd5bbd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.397164] env[62368]: DEBUG nova.network.neutron [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.417728] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198808, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.524963] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.508s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.524963] env[62368]: DEBUG nova.compute.utils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Instance 4785157b-da21-4ed9-aae2-b071b42688e1 could not be found. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 926.524963] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.327s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.527926] env[62368]: INFO nova.compute.claims [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.532753] env[62368]: DEBUG nova.compute.manager [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Instance disappeared during build. {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2513}} [ 926.532753] env[62368]: DEBUG nova.compute.manager [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Unplugging VIFs for instance {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 926.532753] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Acquiring lock "refresh_cache-4785157b-da21-4ed9-aae2-b071b42688e1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.532753] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Acquired lock "refresh_cache-4785157b-da21-4ed9-aae2-b071b42688e1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.532753] env[62368]: DEBUG nova.network.neutron [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 926.612217] env[62368]: DEBUG nova.compute.manager [req-0e4b4452-94e7-4a6b-9a9f-98ced7751e47 req-0be12883-b4da-4cdf-9f1e-0a262f19d82b service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Received event network-vif-deleted-41e7d4b7-38bf-4270-812a-ac72f69e40d6 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.612433] env[62368]: INFO nova.compute.manager [req-0e4b4452-94e7-4a6b-9a9f-98ced7751e47 req-0be12883-b4da-4cdf-9f1e-0a262f19d82b service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Neutron deleted interface 41e7d4b7-38bf-4270-812a-ac72f69e40d6; detaching it from the instance and deleting it from the info cache [ 926.612610] env[62368]: DEBUG nova.network.neutron [req-0e4b4452-94e7-4a6b-9a9f-98ced7751e47 req-0be12883-b4da-4cdf-9f1e-0a262f19d82b service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.828620] env[62368]: INFO nova.compute.manager [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Shelve offloading [ 926.830467] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 926.830755] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b96b857e-f133-46d7-aab7-f59aeb8ea65e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.839664] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 926.839664] env[62368]: value = "task-1198809" [ 926.839664] env[62368]: _type = "Task" [ 926.839664] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.849091] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198809, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.904472] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Releasing lock "refresh_cache-7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.904926] env[62368]: DEBUG nova.compute.manager [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.905202] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 926.909496] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec17f7b-fce3-444b-a097-7bbba2ae52ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.912481] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198808, 'name': ReconfigVM_Task, 'duration_secs': 0.628574} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.912524] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 5633292c-9a74-4c2d-893b-6823f4a587a2/5633292c-9a74-4c2d-893b-6823f4a587a2.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.913547] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87086a01-5bea-440d-bba8-0d304042faf2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.918630] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 926.918788] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37e593f1-b89e-4dc9-9de0-778ada6ca8aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.921707] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 926.921707] env[62368]: value = "task-1198810" [ 926.921707] env[62368]: _type = "Task" [ 926.921707] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.927375] env[62368]: DEBUG oslo_vmware.api [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 926.927375] env[62368]: value = "task-1198811" [ 926.927375] env[62368]: _type = "Task" [ 926.927375] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.935974] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198810, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.942968] env[62368]: DEBUG oslo_vmware.api [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.042122] env[62368]: DEBUG nova.compute.utils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Can not refresh info_cache because instance was not found {{(pid=62368) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 927.080020] env[62368]: DEBUG nova.network.neutron [-] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.115589] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e91ae98-a190-4949-a36a-159c279cd448 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.129638] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306a367e-729f-42fd-a834-297f49ae2f8f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.203028] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.203533] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.205367] env[62368]: DEBUG nova.compute.manager [req-0e4b4452-94e7-4a6b-9a9f-98ced7751e47 req-0be12883-b4da-4cdf-9f1e-0a262f19d82b service nova] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Detach interface failed, port_id=41e7d4b7-38bf-4270-812a-ac72f69e40d6, reason: Instance 14c358a0-434d-472a-a382-b5b0e9bf2bd8 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 927.214595] env[62368]: DEBUG nova.network.neutron [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.290072] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "5af32858-dc9a-4380-827a-daf384c29bed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.290358] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.290579] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "5af32858-dc9a-4380-827a-daf384c29bed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.290765] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.290939] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.293345] env[62368]: INFO nova.compute.manager [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Terminating instance [ 927.295467] env[62368]: DEBUG nova.compute.manager [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 927.295653] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 927.296562] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fae3b00-91f2-4fad-a973-b4e2f7d819fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.305404] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 927.305696] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4a0be95-7428-404a-b7a5-5d851b482997 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.314956] env[62368]: DEBUG oslo_vmware.api [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 927.314956] env[62368]: value = "task-1198812" [ 927.314956] env[62368]: _type = "Task" [ 927.314956] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.331665] env[62368]: DEBUG oslo_vmware.api [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.345591] env[62368]: DEBUG nova.network.neutron [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.356415] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 927.356589] env[62368]: DEBUG nova.compute.manager [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.360665] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1449d6ad-4818-4add-acd6-3322f5bf631d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.368371] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.368371] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.368597] env[62368]: DEBUG nova.network.neutron [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.442951] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198810, 'name': Rename_Task, 'duration_secs': 0.257642} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.443456] env[62368]: DEBUG oslo_vmware.api [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198811, 'name': PowerOffVM_Task, 'duration_secs': 0.23933} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.443757] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.444060] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 927.444236] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 927.444488] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff8e83bf-5711-4c3d-8a83-fb1ff189d43e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.446374] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43b461d8-46dc-48dc-9e25-41cc691465eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.457174] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 927.457174] env[62368]: value = "task-1198815" [ 927.457174] env[62368]: _type = "Task" [ 927.457174] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.491249] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 927.491514] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 927.491728] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Deleting the datastore file [datastore2] 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.492054] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7dffec29-75f7-43c1-b54d-198d0c18524d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.500995] env[62368]: DEBUG oslo_vmware.api [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for the task: (returnval){ [ 927.500995] env[62368]: value = "task-1198816" [ 927.500995] env[62368]: _type = "Task" [ 927.500995] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.512062] env[62368]: DEBUG oslo_vmware.api [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198816, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.584059] env[62368]: INFO nova.compute.manager [-] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Took 1.32 seconds to deallocate network for instance. [ 927.707541] env[62368]: DEBUG nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 927.812682] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf3a309-3229-48e5-b66b-29a77e742f9c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.825692] env[62368]: DEBUG oslo_vmware.api [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198812, 'name': PowerOffVM_Task, 'duration_secs': 0.272479} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.827832] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 927.828085] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 927.828346] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d378bc70-c704-434f-bdec-753cb56c7720 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.830737] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd90607-fb94-44e2-aa43-e207de626d34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.865060] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Releasing lock "refresh_cache-4785157b-da21-4ed9-aae2-b071b42688e1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.865877] env[62368]: DEBUG nova.compute.manager [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62368) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 927.866137] env[62368]: DEBUG nova.compute.manager [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] [instance: 4785157b-da21-4ed9-aae2-b071b42688e1] Skipping network deallocation for instance since networking was not requested. {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2285}} [ 927.873258] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc41083-5084-4665-8505-b3ac01b505ee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.887922] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d245789-5631-4bb7-a6dd-53990e2ec6ff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.908878] env[62368]: DEBUG nova.compute.provider_tree [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.912757] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 927.913055] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 927.913206] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleting the datastore file [datastore2] 5af32858-dc9a-4380-827a-daf384c29bed {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.913775] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-040a0d43-1c5b-46e2-b7ac-f947e99e0a94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.925710] env[62368]: DEBUG oslo_vmware.api [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 927.925710] env[62368]: value = "task-1198818" [ 927.925710] env[62368]: _type = "Task" [ 927.925710] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.938736] env[62368]: DEBUG oslo_vmware.api [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.968717] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198815, 'name': PowerOnVM_Task} progress is 80%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.012846] env[62368]: DEBUG oslo_vmware.api [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Task: {'id': task-1198816, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258596} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.013307] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.013570] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 928.013760] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 928.014050] env[62368]: INFO nova.compute.manager [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Took 1.11 seconds to destroy the instance on the hypervisor. [ 928.014535] env[62368]: DEBUG oslo.service.loopingcall [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.014815] env[62368]: DEBUG nova.compute.manager [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.014924] env[62368]: DEBUG nova.network.neutron [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 928.024968] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.025285] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.056987] env[62368]: DEBUG nova.network.neutron [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 928.097528] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.135980] env[62368]: DEBUG nova.network.neutron [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating instance_info_cache with network_info: [{"id": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "address": "fa:16:3e:d1:c4:2e", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ebe893-6b", "ovs_interfaceid": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.235028] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.415123] env[62368]: DEBUG nova.scheduler.client.report [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.440581] env[62368]: DEBUG oslo_vmware.api [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.420646} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.440581] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.440581] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 928.440581] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 928.440883] env[62368]: INFO nova.compute.manager [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Took 1.14 seconds to destroy the instance on the hypervisor. [ 928.441035] env[62368]: DEBUG oslo.service.loopingcall [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.441164] env[62368]: DEBUG nova.compute.manager [-] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.441257] env[62368]: DEBUG nova.network.neutron [-] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 928.471185] env[62368]: DEBUG oslo_vmware.api [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198815, 'name': PowerOnVM_Task, 'duration_secs': 0.971087} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.471842] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 928.471842] env[62368]: INFO nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Took 10.93 seconds to spawn the instance on the hypervisor. [ 928.471931] env[62368]: DEBUG nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.473065] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9489219-b7d2-44e7-8a11-4f31cc9c018f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.532604] env[62368]: DEBUG nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 928.559958] env[62368]: DEBUG nova.network.neutron [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.641063] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.888543] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e3148785-1479-45e7-a083-03e7550f60bf tempest-ServersListShow296Test-138057209 tempest-ServersListShow296Test-138057209-project-member] Lock "4785157b-da21-4ed9-aae2-b071b42688e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 3.894s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.911289] env[62368]: DEBUG nova.compute.manager [req-d6241aac-1abc-4f6a-a32c-388ea3a4ba85 req-2838ab0c-d93e-4cdd-aefc-cea47d40cee2 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Received event network-vif-deleted-8be1ff35-0e69-42f1-a582-16bf1f496a2d {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.911289] env[62368]: INFO nova.compute.manager [req-d6241aac-1abc-4f6a-a32c-388ea3a4ba85 req-2838ab0c-d93e-4cdd-aefc-cea47d40cee2 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Neutron deleted interface 8be1ff35-0e69-42f1-a582-16bf1f496a2d; detaching it from the instance and deleting it from the info cache [ 928.911289] env[62368]: DEBUG nova.network.neutron [req-d6241aac-1abc-4f6a-a32c-388ea3a4ba85 req-2838ab0c-d93e-4cdd-aefc-cea47d40cee2 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.927021] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.927021] env[62368]: DEBUG nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 928.930233] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.832s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.930706] env[62368]: DEBUG nova.objects.instance [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lazy-loading 'resources' on Instance uuid 14c358a0-434d-472a-a382-b5b0e9bf2bd8 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.996524] env[62368]: INFO nova.compute.manager [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Took 25.22 seconds to build instance. [ 929.009326] env[62368]: DEBUG nova.compute.manager [req-c1f78baa-3c8a-4bbf-b23a-0ec7bc068b25 req-469f8784-31c5-4fde-975b-3c9ae64d2808 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received event network-vif-unplugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.010030] env[62368]: DEBUG oslo_concurrency.lockutils [req-c1f78baa-3c8a-4bbf-b23a-0ec7bc068b25 req-469f8784-31c5-4fde-975b-3c9ae64d2808 service nova] Acquiring lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.010257] env[62368]: DEBUG oslo_concurrency.lockutils [req-c1f78baa-3c8a-4bbf-b23a-0ec7bc068b25 req-469f8784-31c5-4fde-975b-3c9ae64d2808 service nova] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.010437] env[62368]: DEBUG oslo_concurrency.lockutils [req-c1f78baa-3c8a-4bbf-b23a-0ec7bc068b25 req-469f8784-31c5-4fde-975b-3c9ae64d2808 service nova] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.010618] env[62368]: DEBUG nova.compute.manager [req-c1f78baa-3c8a-4bbf-b23a-0ec7bc068b25 req-469f8784-31c5-4fde-975b-3c9ae64d2808 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] No waiting events found dispatching network-vif-unplugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.010850] env[62368]: WARNING nova.compute.manager [req-c1f78baa-3c8a-4bbf-b23a-0ec7bc068b25 req-469f8784-31c5-4fde-975b-3c9ae64d2808 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received unexpected event network-vif-unplugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e for instance with vm_state shelved and task_state shelving_offloading. [ 929.031078] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 929.032037] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae587e94-674d-40e9-b28d-173ad8ed74fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.045333] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 929.045594] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a728f9f0-7340-447b-9dee-a0d2b30f965a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.052738] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.064026] env[62368]: INFO nova.compute.manager [-] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Took 1.05 seconds to deallocate network for instance. [ 929.144235] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 929.144493] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 929.144751] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleting the datastore file [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.145076] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2db9a00-2f99-42ad-8c2f-4dbfec0b3202 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.165048] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 929.165048] env[62368]: value = "task-1198820" [ 929.165048] env[62368]: _type = "Task" [ 929.165048] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.178366] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198820, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.383153] env[62368]: DEBUG nova.network.neutron [-] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.415145] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e93a3fab-997f-411e-bbb8-38549f14e377 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.428057] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709f00ee-053d-4724-85a8-a48d83d3b41b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.440979] env[62368]: DEBUG nova.compute.utils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.446926] env[62368]: DEBUG nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 929.448630] env[62368]: DEBUG nova.network.neutron [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 929.476139] env[62368]: DEBUG nova.compute.manager [req-d6241aac-1abc-4f6a-a32c-388ea3a4ba85 req-2838ab0c-d93e-4cdd-aefc-cea47d40cee2 service nova] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Detach interface failed, port_id=8be1ff35-0e69-42f1-a582-16bf1f496a2d, reason: Instance 5af32858-dc9a-4380-827a-daf384c29bed could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 929.496678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f8a183f-989a-4a99-a124-db5a78950088 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.734s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.536509] env[62368]: DEBUG nova.policy [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c66c425264424e3eaa37c8058e62cb30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '026a3176abc34001ab3506445226777f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 929.571477] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.679577] env[62368]: DEBUG oslo_vmware.api [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198820, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311317} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.683910] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.684659] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 929.685924] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 929.721807] env[62368]: INFO nova.scheduler.client.report [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleted allocations for instance eea21546-fbbf-4440-829c-8583c4ccabb6 [ 929.786691] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47c8a41-61d8-4877-a752-b853ca504dfa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.794157] env[62368]: DEBUG oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5201ba4f-ea86-999f-6868-89acad5b0a80/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 929.794157] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2a745f-b254-4f5e-ae18-c16ee8b84a91 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.803415] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d7cf85-d879-4666-a0d4-c29db0da4630 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.808481] env[62368]: DEBUG oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5201ba4f-ea86-999f-6868-89acad5b0a80/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 929.812292] env[62368]: ERROR oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5201ba4f-ea86-999f-6868-89acad5b0a80/disk-0.vmdk due to incomplete transfer. [ 929.815439] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ba483756-e47c-40e9-9f29-79783e0c3585 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.846406] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3032d1ee-1e01-4e35-8207-af105f53e3d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.849331] env[62368]: DEBUG oslo_vmware.rw_handles [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5201ba4f-ea86-999f-6868-89acad5b0a80/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 929.849475] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Uploaded image c4f6a28c-a893-4200-ba2f-b1e0d47d22df to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 929.851924] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 929.852188] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a4d0af79-037b-49d8-8fc2-d7d788cc011f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.860610] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029f0fd4-0a87-47c0-8ea8-9cce2569aef0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.865216] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 929.865216] env[62368]: value = "task-1198822" [ 929.865216] env[62368]: _type = "Task" [ 929.865216] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.877055] env[62368]: DEBUG nova.compute.provider_tree [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.883844] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198822, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.887110] env[62368]: INFO nova.compute.manager [-] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Took 1.45 seconds to deallocate network for instance. [ 929.949651] env[62368]: DEBUG nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 930.113957] env[62368]: DEBUG nova.network.neutron [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Successfully created port: 318e6d00-8a0d-4405-b405-50d208bc29f8 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.227365] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.379180] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198822, 'name': Destroy_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.380322] env[62368]: DEBUG nova.scheduler.client.report [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.396715] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.879860] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198822, 'name': Destroy_Task, 'duration_secs': 0.952347} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.880379] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Destroyed the VM [ 930.880648] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 930.880925] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-483f4d4a-b756-4fe2-9ebd-64ec7396367e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.886583] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.957s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.893466] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.659s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.896382] env[62368]: INFO nova.compute.claims [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.903890] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 930.903890] env[62368]: value = "task-1198823" [ 930.903890] env[62368]: _type = "Task" [ 930.903890] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.916442] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198823, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.927395] env[62368]: INFO nova.scheduler.client.report [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted allocations for instance 14c358a0-434d-472a-a382-b5b0e9bf2bd8 [ 930.962686] env[62368]: DEBUG nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 930.997414] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.997658] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.997878] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.002523] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.002523] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.002718] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.002965] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.003440] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.003729] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.003957] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.004213] env[62368]: DEBUG nova.virt.hardware [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.006150] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd90c2a-da04-4233-9577-5ff88acfe990 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.016131] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0786225c-6f43-4401-89df-49291949acdc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.176739] env[62368]: DEBUG nova.compute.manager [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received event network-changed-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.176943] env[62368]: DEBUG nova.compute.manager [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Refreshing instance network info cache due to event network-changed-b2ebe893-6ba3-40da-8def-fb6cf1ab442e. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.178062] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Acquiring lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.178062] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Acquired lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.178062] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Refreshing network info cache for port b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 931.419932] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198823, 'name': RemoveSnapshot_Task, 'duration_secs': 0.459514} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.421218] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 931.421495] env[62368]: DEBUG nova.compute.manager [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.424074] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548046d5-0250-4485-8513-e8f39367ab4d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.439972] env[62368]: DEBUG oslo_concurrency.lockutils [None req-34ce656f-cc6e-424b-9fdb-d3e907964b63 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "14c358a0-434d-472a-a382-b5b0e9bf2bd8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.805s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.638837] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "eea21546-fbbf-4440-829c-8583c4ccabb6" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.864534] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.864534] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.942155] env[62368]: INFO nova.compute.manager [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Shelve offloading [ 931.947916] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 931.947916] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efb0e194-b105-423a-864a-99d64fb79e26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.956299] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 931.956299] env[62368]: value = "task-1198825" [ 931.956299] env[62368]: _type = "Task" [ 931.956299] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.965824] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198825, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.152271] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updated VIF entry in instance network info cache for port b2ebe893-6ba3-40da-8def-fb6cf1ab442e. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 932.152627] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating instance_info_cache with network_info: [{"id": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "address": "fa:16:3e:d1:c4:2e", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb2ebe893-6b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.175083] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0730fde2-83a8-4e3c-b994-863808db3a5e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.183823] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a95bde-121f-4efb-805c-89086d72ea79 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.218052] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c593f268-7756-4338-bfcb-25eaadc4c422 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.232986] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d119ba51-4e16-4021-b8c7-e87e671245e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.248571] env[62368]: DEBUG nova.compute.provider_tree [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.366124] env[62368]: DEBUG nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.468604] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 932.468818] env[62368]: DEBUG nova.compute.manager [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.469603] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2e5feb-ee88-4472-bf5c-7defb93d313c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.475989] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.476183] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.476363] env[62368]: DEBUG nova.network.neutron [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 932.499802] env[62368]: DEBUG nova.network.neutron [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Successfully updated port: 318e6d00-8a0d-4405-b405-50d208bc29f8 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.607014] env[62368]: DEBUG nova.compute.manager [req-ee9cc883-1453-4a3e-b2d4-6067fb299496 req-fdcac261-c6d6-4dd5-8b59-9ba7f829aefc service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received event network-changed-14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.607261] env[62368]: DEBUG nova.compute.manager [req-ee9cc883-1453-4a3e-b2d4-6067fb299496 req-fdcac261-c6d6-4dd5-8b59-9ba7f829aefc service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing instance network info cache due to event network-changed-14840bf2-7da7-4d91-aa96-02a1771fb7d1. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 932.607430] env[62368]: DEBUG oslo_concurrency.lockutils [req-ee9cc883-1453-4a3e-b2d4-6067fb299496 req-fdcac261-c6d6-4dd5-8b59-9ba7f829aefc service nova] Acquiring lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.607561] env[62368]: DEBUG oslo_concurrency.lockutils [req-ee9cc883-1453-4a3e-b2d4-6067fb299496 req-fdcac261-c6d6-4dd5-8b59-9ba7f829aefc service nova] Acquired lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.607761] env[62368]: DEBUG nova.network.neutron [req-ee9cc883-1453-4a3e-b2d4-6067fb299496 req-fdcac261-c6d6-4dd5-8b59-9ba7f829aefc service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing network info cache for port 14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.656713] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Releasing lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.656713] env[62368]: DEBUG nova.compute.manager [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.656713] env[62368]: DEBUG nova.compute.manager [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing instance network info cache due to event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 932.656713] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.656713] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.656713] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.751694] env[62368]: DEBUG nova.scheduler.client.report [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.897364] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.002372] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "refresh_cache-e427d283-1f8f-4c7a-869c-4cd48a595470" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.002656] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "refresh_cache-e427d283-1f8f-4c7a-869c-4cd48a595470" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.002656] env[62368]: DEBUG nova.network.neutron [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.257155] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.257576] env[62368]: DEBUG nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.263193] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.210s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.264176] env[62368]: INFO nova.compute.claims [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.331699] env[62368]: DEBUG nova.network.neutron [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Updating instance_info_cache with network_info: [{"id": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "address": "fa:16:3e:17:f6:62", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58072bc6-80", "ovs_interfaceid": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.352432] env[62368]: DEBUG nova.compute.manager [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Received event network-vif-plugged-318e6d00-8a0d-4405-b405-50d208bc29f8 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.352679] env[62368]: DEBUG oslo_concurrency.lockutils [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] Acquiring lock "e427d283-1f8f-4c7a-869c-4cd48a595470-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.352897] env[62368]: DEBUG oslo_concurrency.lockutils [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] Lock "e427d283-1f8f-4c7a-869c-4cd48a595470-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.353061] env[62368]: DEBUG oslo_concurrency.lockutils [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] Lock "e427d283-1f8f-4c7a-869c-4cd48a595470-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.353347] env[62368]: DEBUG nova.compute.manager [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] No waiting events found dispatching network-vif-plugged-318e6d00-8a0d-4405-b405-50d208bc29f8 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.353621] env[62368]: WARNING nova.compute.manager [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Received unexpected event network-vif-plugged-318e6d00-8a0d-4405-b405-50d208bc29f8 for instance with vm_state building and task_state spawning. [ 933.353803] env[62368]: DEBUG nova.compute.manager [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Received event network-changed-318e6d00-8a0d-4405-b405-50d208bc29f8 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.353977] env[62368]: DEBUG nova.compute.manager [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Refreshing instance network info cache due to event network-changed-318e6d00-8a0d-4405-b405-50d208bc29f8. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.354177] env[62368]: DEBUG oslo_concurrency.lockutils [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] Acquiring lock "refresh_cache-e427d283-1f8f-4c7a-869c-4cd48a595470" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.543687] env[62368]: DEBUG nova.network.neutron [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 933.575175] env[62368]: DEBUG nova.network.neutron [req-ee9cc883-1453-4a3e-b2d4-6067fb299496 req-fdcac261-c6d6-4dd5-8b59-9ba7f829aefc service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updated VIF entry in instance network info cache for port 14840bf2-7da7-4d91-aa96-02a1771fb7d1. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 933.575175] env[62368]: DEBUG nova.network.neutron [req-ee9cc883-1453-4a3e-b2d4-6067fb299496 req-fdcac261-c6d6-4dd5-8b59-9ba7f829aefc service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.616374] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updated VIF entry in instance network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 933.616729] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.699905] env[62368]: DEBUG nova.network.neutron [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Updating instance_info_cache with network_info: [{"id": "318e6d00-8a0d-4405-b405-50d208bc29f8", "address": "fa:16:3e:b1:56:8e", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318e6d00-8a", "ovs_interfaceid": "318e6d00-8a0d-4405-b405-50d208bc29f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.769486] env[62368]: DEBUG nova.compute.utils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.771693] env[62368]: DEBUG nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.772131] env[62368]: DEBUG nova.network.neutron [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 933.815582] env[62368]: DEBUG nova.policy [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50e00b1a1fd3401fa803dea5b0801965', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73819709966b4e5f873020b399c49871', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.833857] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.038289] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "1d5a93f0-1068-4514-b01c-011e95f48498" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.038839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.077855] env[62368]: DEBUG oslo_concurrency.lockutils [req-ee9cc883-1453-4a3e-b2d4-6067fb299496 req-fdcac261-c6d6-4dd5-8b59-9ba7f829aefc service nova] Releasing lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.090567] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 934.091587] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e774824e-5176-461e-ba88-b5c1a6904ae0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.103250] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 934.103657] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d625dc3-ff07-4586-87db-f59b74e85e30 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.119522] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.119875] env[62368]: DEBUG nova.compute.manager [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received event network-changed-14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.120112] env[62368]: DEBUG nova.compute.manager [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing instance network info cache due to event network-changed-14840bf2-7da7-4d91-aa96-02a1771fb7d1. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.121316] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Acquiring lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.121316] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Acquired lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.121316] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing network info cache for port 14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 934.134136] env[62368]: DEBUG nova.network.neutron [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Successfully created port: 05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.202418] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "refresh_cache-e427d283-1f8f-4c7a-869c-4cd48a595470" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.202758] env[62368]: DEBUG nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Instance network_info: |[{"id": "318e6d00-8a0d-4405-b405-50d208bc29f8", "address": "fa:16:3e:b1:56:8e", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318e6d00-8a", "ovs_interfaceid": "318e6d00-8a0d-4405-b405-50d208bc29f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 934.203104] env[62368]: DEBUG oslo_concurrency.lockutils [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] Acquired lock "refresh_cache-e427d283-1f8f-4c7a-869c-4cd48a595470" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.203354] env[62368]: DEBUG nova.network.neutron [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Refreshing network info cache for port 318e6d00-8a0d-4405-b405-50d208bc29f8 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 934.209701] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:56:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8f40f5c4-c146-449c-884d-6f884dcf2acf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '318e6d00-8a0d-4405-b405-50d208bc29f8', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.216901] env[62368]: DEBUG oslo.service.loopingcall [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.220700] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 934.221382] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ac1e0d8-b056-4838-896f-be466149236c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.244920] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.244920] env[62368]: value = "task-1198828" [ 934.244920] env[62368]: _type = "Task" [ 934.244920] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.254255] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198828, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.272986] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 934.273267] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 934.273461] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleting the datastore file [datastore2] 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.273733] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6bd0d3c8-95b9-4b28-b1b3-b1e6a8415bac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.278019] env[62368]: DEBUG nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.285309] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 934.285309] env[62368]: value = "task-1198829" [ 934.285309] env[62368]: _type = "Task" [ 934.285309] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.293694] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198829, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.545601] env[62368]: DEBUG nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.659257] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a34805-11c9-42bb-81e5-5e3be2215c2d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.667937] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903c413e-40fa-4f18-9725-b94ae00ace65 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.707695] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227175a0-fd60-43ea-a8ed-3690bd42a87c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.713738] env[62368]: DEBUG nova.compute.manager [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.713964] env[62368]: DEBUG nova.compute.manager [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing instance network info cache due to event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.714229] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.714385] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.714581] env[62368]: DEBUG nova.network.neutron [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 934.720409] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7291f13d-6d73-4e4e-8643-6d33015503ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.737296] env[62368]: DEBUG nova.compute.provider_tree [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.756357] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198828, 'name': CreateVM_Task, 'duration_secs': 0.412725} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.756357] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 934.756357] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.756357] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.756357] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.756357] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13d93985-b34c-4efd-9e9e-6ffa18ad9f2a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.762465] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 934.762465] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52233f35-a9f4-3ae3-c1e5-105a6b9793e5" [ 934.762465] env[62368]: _type = "Task" [ 934.762465] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.773057] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52233f35-a9f4-3ae3-c1e5-105a6b9793e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.799029] env[62368]: DEBUG oslo_vmware.api [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198829, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236278} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.799029] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.799029] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 934.799029] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 934.819711] env[62368]: INFO nova.scheduler.client.report [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted allocations for instance 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e [ 934.834866] env[62368]: DEBUG nova.network.neutron [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Updated VIF entry in instance network info cache for port 318e6d00-8a0d-4405-b405-50d208bc29f8. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 934.835291] env[62368]: DEBUG nova.network.neutron [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Updating instance_info_cache with network_info: [{"id": "318e6d00-8a0d-4405-b405-50d208bc29f8", "address": "fa:16:3e:b1:56:8e", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318e6d00-8a", "ovs_interfaceid": "318e6d00-8a0d-4405-b405-50d208bc29f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.966195] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updated VIF entry in instance network info cache for port 14840bf2-7da7-4d91-aa96-02a1771fb7d1. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 934.966638] env[62368]: DEBUG nova.network.neutron [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.067449] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.242806] env[62368]: DEBUG nova.scheduler.client.report [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.274723] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52233f35-a9f4-3ae3-c1e5-105a6b9793e5, 'name': SearchDatastore_Task, 'duration_secs': 0.019039} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.275106] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.275417] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.275661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.275812] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.275992] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.276545] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-241739a8-1780-449f-848f-d30b2cc1827f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.285908] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.286108] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 935.286830] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d588555-40be-46e7-a50c-27b4eafee97e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.289958] env[62368]: DEBUG nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.295285] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 935.295285] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]528a99c5-32d5-f483-a086-4c533174cb51" [ 935.295285] env[62368]: _type = "Task" [ 935.295285] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.303419] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528a99c5-32d5-f483-a086-4c533174cb51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.315130] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.315426] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.315596] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.315783] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.315932] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.316093] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.316304] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.316477] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.316648] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.316813] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.316986] env[62368]: DEBUG nova.virt.hardware [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.317915] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d8157f-9244-470b-9c4c-2b46d461213f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.327348] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.328894] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4786b8da-af11-4ed1-810f-8d262d8c94c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.344556] env[62368]: DEBUG oslo_concurrency.lockutils [req-810a1d98-d7cd-4eb6-8451-e4b699b107d1 req-98b35b1e-4537-431a-9f05-0a75965cdf60 service nova] Releasing lock "refresh_cache-e427d283-1f8f-4c7a-869c-4cd48a595470" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.469528] env[62368]: DEBUG oslo_concurrency.lockutils [req-876e21c2-fb16-4ed3-9425-b4e3146c3882 req-1fbf90c2-08f9-47c9-85be-595b76eab7d6 service nova] Releasing lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.493469] env[62368]: DEBUG nova.network.neutron [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updated VIF entry in instance network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 935.493878] env[62368]: DEBUG nova.network.neutron [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.689188] env[62368]: DEBUG nova.compute.manager [req-5ccba1c6-0911-4c3c-98f3-2b93c7a61bb3 req-b95f346a-30b2-4b31-a91a-b7eb86bb37ac service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Received event network-vif-plugged-05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.689552] env[62368]: DEBUG oslo_concurrency.lockutils [req-5ccba1c6-0911-4c3c-98f3-2b93c7a61bb3 req-b95f346a-30b2-4b31-a91a-b7eb86bb37ac service nova] Acquiring lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.690918] env[62368]: DEBUG oslo_concurrency.lockutils [req-5ccba1c6-0911-4c3c-98f3-2b93c7a61bb3 req-b95f346a-30b2-4b31-a91a-b7eb86bb37ac service nova] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.691251] env[62368]: DEBUG oslo_concurrency.lockutils [req-5ccba1c6-0911-4c3c-98f3-2b93c7a61bb3 req-b95f346a-30b2-4b31-a91a-b7eb86bb37ac service nova] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.691544] env[62368]: DEBUG nova.compute.manager [req-5ccba1c6-0911-4c3c-98f3-2b93c7a61bb3 req-b95f346a-30b2-4b31-a91a-b7eb86bb37ac service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] No waiting events found dispatching network-vif-plugged-05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.691840] env[62368]: WARNING nova.compute.manager [req-5ccba1c6-0911-4c3c-98f3-2b93c7a61bb3 req-b95f346a-30b2-4b31-a91a-b7eb86bb37ac service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Received unexpected event network-vif-plugged-05119e53-57b5-4b7d-a056-3bb898352179 for instance with vm_state building and task_state spawning. [ 935.748818] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.749420] env[62368]: DEBUG nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 935.752014] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.181s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.752366] env[62368]: DEBUG nova.objects.instance [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lazy-loading 'resources' on Instance uuid 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.791712] env[62368]: DEBUG nova.network.neutron [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Successfully updated port: 05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.806124] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528a99c5-32d5-f483-a086-4c533174cb51, 'name': SearchDatastore_Task, 'duration_secs': 0.0214} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.807461] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19d1b4ce-f6e3-4d94-be9f-c84412e448f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.813740] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 935.813740] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524801bd-d39c-4572-ab66-7858a77b61c2" [ 935.813740] env[62368]: _type = "Task" [ 935.813740] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.822196] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524801bd-d39c-4572-ab66-7858a77b61c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.997777] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.998025] env[62368]: DEBUG nova.compute.manager [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Received event network-vif-unplugged-58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.998240] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Acquiring lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.998477] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.998646] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.998812] env[62368]: DEBUG nova.compute.manager [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] No waiting events found dispatching network-vif-unplugged-58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.998985] env[62368]: WARNING nova.compute.manager [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Received unexpected event network-vif-unplugged-58072bc6-80bf-4d1f-b471-2ecdd115a8ad for instance with vm_state shelved and task_state shelving_offloading. [ 935.999173] env[62368]: DEBUG nova.compute.manager [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Received event network-changed-58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.999364] env[62368]: DEBUG nova.compute.manager [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Refreshing instance network info cache due to event network-changed-58072bc6-80bf-4d1f-b471-2ecdd115a8ad. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.999560] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Acquiring lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.999700] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Acquired lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.999858] env[62368]: DEBUG nova.network.neutron [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Refreshing network info cache for port 58072bc6-80bf-4d1f-b471-2ecdd115a8ad {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 936.256058] env[62368]: DEBUG nova.compute.utils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 936.260896] env[62368]: DEBUG nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 936.260896] env[62368]: DEBUG nova.network.neutron [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 936.299831] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.299831] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.299831] env[62368]: DEBUG nova.network.neutron [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 936.309009] env[62368]: DEBUG nova.policy [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0800ab273ca04fbf9396175b57eed6b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e821059910b4e32aab596c6f4d521d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.325292] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524801bd-d39c-4572-ab66-7858a77b61c2, 'name': SearchDatastore_Task, 'duration_secs': 0.035503} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.325564] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.325819] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] e427d283-1f8f-4c7a-869c-4cd48a595470/e427d283-1f8f-4c7a-869c-4cd48a595470.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 936.326090] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46212582-10e5-4ae4-b545-7965bd591694 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.333266] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 936.333266] env[62368]: value = "task-1198831" [ 936.333266] env[62368]: _type = "Task" [ 936.333266] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.345981] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.478496] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1465de13-cac3-4179-b1ad-3649eae0095a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.486767] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f83657f-7ada-45ea-a16a-4b1f9ea9d3db {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.524184] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c95088f-e1da-4971-8cd9-d3bc27b1dfba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.533061] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add24263-7f21-4ec7-bf03-ecf0db3b8ada {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.548375] env[62368]: DEBUG nova.compute.provider_tree [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.562833] env[62368]: DEBUG nova.network.neutron [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Successfully created port: d229b53b-0e94-41c9-b8e9-0b0eaa366d3c {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.596696] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.761998] env[62368]: DEBUG nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 936.771707] env[62368]: DEBUG nova.network.neutron [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Updated VIF entry in instance network info cache for port 58072bc6-80bf-4d1f-b471-2ecdd115a8ad. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.773855] env[62368]: DEBUG nova.network.neutron [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Updating instance_info_cache with network_info: [{"id": "58072bc6-80bf-4d1f-b471-2ecdd115a8ad", "address": "fa:16:3e:17:f6:62", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": null, "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap58072bc6-80", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.834088] env[62368]: DEBUG nova.network.neutron [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 936.847655] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510864} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.848285] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] e427d283-1f8f-4c7a-869c-4cd48a595470/e427d283-1f8f-4c7a-869c-4cd48a595470.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.848859] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.849169] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8553bfac-a592-4e7e-a0a6-3759add3020c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.856965] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 936.856965] env[62368]: value = "task-1198832" [ 936.856965] env[62368]: _type = "Task" [ 936.856965] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.865308] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.981584] env[62368]: DEBUG nova.network.neutron [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [{"id": "05119e53-57b5-4b7d-a056-3bb898352179", "address": "fa:16:3e:3b:b9:de", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05119e53-57", "ovs_interfaceid": "05119e53-57b5-4b7d-a056-3bb898352179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.051756] env[62368]: DEBUG nova.scheduler.client.report [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.276151] env[62368]: DEBUG oslo_concurrency.lockutils [req-1862b08f-a8a4-4a3f-a516-19af70773c6c req-7d3afe98-5e35-4c34-8011-7bb790c2b097 service nova] Releasing lock "refresh_cache-3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.366516] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061158} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.366859] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.367549] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3f2be4-5dfe-4160-9022-f840154b6d10 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.389558] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] e427d283-1f8f-4c7a-869c-4cd48a595470/e427d283-1f8f-4c7a-869c-4cd48a595470.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.389819] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5edad46c-6cfb-487b-a2e0-d8aea3672f26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.418091] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 937.418091] env[62368]: value = "task-1198833" [ 937.418091] env[62368]: _type = "Task" [ 937.418091] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.429939] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198833, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.484431] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.484631] env[62368]: DEBUG nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Instance network_info: |[{"id": "05119e53-57b5-4b7d-a056-3bb898352179", "address": "fa:16:3e:3b:b9:de", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05119e53-57", "ovs_interfaceid": "05119e53-57b5-4b7d-a056-3bb898352179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.485081] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:b9:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0685bd0b-3dbf-4a06-951c-c6a4726dd4b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05119e53-57b5-4b7d-a056-3bb898352179', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.493229] env[62368]: DEBUG oslo.service.loopingcall [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.493527] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 937.493785] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cea45d20-b7f2-4a09-9f3e-6c267c7bc6a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.514615] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.514615] env[62368]: value = "task-1198834" [ 937.514615] env[62368]: _type = "Task" [ 937.514615] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.522782] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198834, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.557131] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.805s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.560252] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.333s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.560597] env[62368]: DEBUG nova.objects.instance [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lazy-loading 'resources' on Instance uuid eea21546-fbbf-4440-829c-8583c4ccabb6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.582811] env[62368]: INFO nova.scheduler.client.report [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Deleted allocations for instance 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc [ 937.719768] env[62368]: DEBUG nova.compute.manager [req-c4d99aa2-a271-44b3-89c0-63aabedfe27a req-ecf08504-fc9f-4aa2-829f-673b67cae09b service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Received event network-changed-05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.719768] env[62368]: DEBUG nova.compute.manager [req-c4d99aa2-a271-44b3-89c0-63aabedfe27a req-ecf08504-fc9f-4aa2-829f-673b67cae09b service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Refreshing instance network info cache due to event network-changed-05119e53-57b5-4b7d-a056-3bb898352179. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.719768] env[62368]: DEBUG oslo_concurrency.lockutils [req-c4d99aa2-a271-44b3-89c0-63aabedfe27a req-ecf08504-fc9f-4aa2-829f-673b67cae09b service nova] Acquiring lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.719768] env[62368]: DEBUG oslo_concurrency.lockutils [req-c4d99aa2-a271-44b3-89c0-63aabedfe27a req-ecf08504-fc9f-4aa2-829f-673b67cae09b service nova] Acquired lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.719768] env[62368]: DEBUG nova.network.neutron [req-c4d99aa2-a271-44b3-89c0-63aabedfe27a req-ecf08504-fc9f-4aa2-829f-673b67cae09b service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Refreshing network info cache for port 05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.771941] env[62368]: DEBUG nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 937.802403] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.802719] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.802921] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.803216] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.803435] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.803641] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.803896] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.804112] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.804328] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.804542] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.804763] env[62368]: DEBUG nova.virt.hardware [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.805758] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b6d2f9-fb9f-442b-b3fa-d0642fb40696 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.814889] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120e1360-2f32-48f2-80b3-6d8ad1278ed1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.928595] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198833, 'name': ReconfigVM_Task, 'duration_secs': 0.287664} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.929038] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Reconfigured VM instance instance-0000005b to attach disk [datastore2] e427d283-1f8f-4c7a-869c-4cd48a595470/e427d283-1f8f-4c7a-869c-4cd48a595470.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.929562] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f36e797-f793-42cc-9b5b-e8a55c018d7f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.938614] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 937.938614] env[62368]: value = "task-1198836" [ 937.938614] env[62368]: _type = "Task" [ 937.938614] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.947967] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198836, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.026615] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198834, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.063652] env[62368]: DEBUG nova.objects.instance [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lazy-loading 'numa_topology' on Instance uuid eea21546-fbbf-4440-829c-8583c4ccabb6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.096304] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8959a0e3-ebb3-4f1f-b0b5-2a37be90d4c7 tempest-ServerShowV254Test-1361232689 tempest-ServerShowV254Test-1361232689-project-member] Lock "7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.315s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.385294] env[62368]: DEBUG nova.network.neutron [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Successfully updated port: d229b53b-0e94-41c9-b8e9-0b0eaa366d3c {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.450848] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198836, 'name': Rename_Task, 'duration_secs': 0.164289} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.451056] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 938.451180] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd65b991-79ac-4f34-941a-7519f70f5d40 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.458766] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 938.458766] env[62368]: value = "task-1198837" [ 938.458766] env[62368]: _type = "Task" [ 938.458766] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.467530] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198837, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.482284] env[62368]: DEBUG nova.network.neutron [req-c4d99aa2-a271-44b3-89c0-63aabedfe27a req-ecf08504-fc9f-4aa2-829f-673b67cae09b service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updated VIF entry in instance network info cache for port 05119e53-57b5-4b7d-a056-3bb898352179. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.482653] env[62368]: DEBUG nova.network.neutron [req-c4d99aa2-a271-44b3-89c0-63aabedfe27a req-ecf08504-fc9f-4aa2-829f-673b67cae09b service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [{"id": "05119e53-57b5-4b7d-a056-3bb898352179", "address": "fa:16:3e:3b:b9:de", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05119e53-57", "ovs_interfaceid": "05119e53-57b5-4b7d-a056-3bb898352179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.535489] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198834, 'name': CreateVM_Task, 'duration_secs': 0.7346} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.535489] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 938.535489] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.535489] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.535489] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.535489] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48d1917b-a660-44fb-b9bb-b956de6bc74b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.540844] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 938.540844] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52aa0627-6cfa-7f80-5269-ce08ccb04630" [ 938.540844] env[62368]: _type = "Task" [ 938.540844] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.550131] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52aa0627-6cfa-7f80-5269-ce08ccb04630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.567110] env[62368]: DEBUG nova.objects.base [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 938.752080] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee157d37-1824-4332-9764-d348f268407f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.761121] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684d2669-f8cc-4b0d-b66b-0340f42ac888 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.793313] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1f370e-d58a-46fa-9be7-f527f3f3c61d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.801347] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b03075-91e2-453b-85e8-b82cdd81fce3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.815366] env[62368]: DEBUG nova.compute.provider_tree [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.887978] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.888177] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.888339] env[62368]: DEBUG nova.network.neutron [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 938.936039] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "interface-a5cbadbd-20dd-4514-8867-20243af5db0c-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.936342] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-a5cbadbd-20dd-4514-8867-20243af5db0c-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.936734] env[62368]: DEBUG nova.objects.instance [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'flavor' on Instance uuid a5cbadbd-20dd-4514-8867-20243af5db0c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.968897] env[62368]: DEBUG oslo_vmware.api [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198837, 'name': PowerOnVM_Task, 'duration_secs': 0.495406} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.969045] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.969259] env[62368]: INFO nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Took 8.01 seconds to spawn the instance on the hypervisor. [ 938.969443] env[62368]: DEBUG nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.970250] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088f2185-91fb-45f9-9198-72eae61845fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.984633] env[62368]: DEBUG oslo_concurrency.lockutils [req-c4d99aa2-a271-44b3-89c0-63aabedfe27a req-ecf08504-fc9f-4aa2-829f-673b67cae09b service nova] Releasing lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.051843] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52aa0627-6cfa-7f80-5269-ce08ccb04630, 'name': SearchDatastore_Task, 'duration_secs': 0.010658} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.052156] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.052394] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.052658] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.052816] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.053008] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.053281] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbbac528-a03a-4be5-b61f-58184a0bb069 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.063029] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.063029] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 939.063029] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f54ccec6-73c3-4f44-99c7-6768f273d7b4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.068499] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 939.068499] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]528809a1-c401-95c5-3ba8-f7781f1e315f" [ 939.068499] env[62368]: _type = "Task" [ 939.068499] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.076486] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528809a1-c401-95c5-3ba8-f7781f1e315f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.318141] env[62368]: DEBUG nova.scheduler.client.report [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.431138] env[62368]: DEBUG nova.network.neutron [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 939.489510] env[62368]: INFO nova.compute.manager [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Took 13.31 seconds to build instance. [ 939.529835] env[62368]: DEBUG nova.objects.instance [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'pci_requests' on Instance uuid a5cbadbd-20dd-4514-8867-20243af5db0c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.579739] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528809a1-c401-95c5-3ba8-f7781f1e315f, 'name': SearchDatastore_Task, 'duration_secs': 0.0095} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.580555] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-682e39ce-7760-4b80-9b1c-04641fef410c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.585766] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 939.585766] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52da14bc-8389-f15b-ac31-006fae86077d" [ 939.585766] env[62368]: _type = "Task" [ 939.585766] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.594229] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52da14bc-8389-f15b-ac31-006fae86077d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.633494] env[62368]: DEBUG nova.network.neutron [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance_info_cache with network_info: [{"id": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "address": "fa:16:3e:b0:23:0b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd229b53b-0e", "ovs_interfaceid": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.823501] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.263s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.826503] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.430s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.826630] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.829382] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.932s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.833020] env[62368]: INFO nova.compute.claims [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.854980] env[62368]: INFO nova.scheduler.client.report [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted allocations for instance 5af32858-dc9a-4380-827a-daf384c29bed [ 939.938549] env[62368]: DEBUG nova.compute.manager [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Received event network-vif-plugged-d229b53b-0e94-41c9-b8e9-0b0eaa366d3c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.938549] env[62368]: DEBUG oslo_concurrency.lockutils [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] Acquiring lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.938652] env[62368]: DEBUG oslo_concurrency.lockutils [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.938737] env[62368]: DEBUG oslo_concurrency.lockutils [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.938912] env[62368]: DEBUG nova.compute.manager [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] No waiting events found dispatching network-vif-plugged-d229b53b-0e94-41c9-b8e9-0b0eaa366d3c {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.939278] env[62368]: WARNING nova.compute.manager [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Received unexpected event network-vif-plugged-d229b53b-0e94-41c9-b8e9-0b0eaa366d3c for instance with vm_state building and task_state spawning. [ 939.939508] env[62368]: DEBUG nova.compute.manager [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Received event network-changed-d229b53b-0e94-41c9-b8e9-0b0eaa366d3c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.939677] env[62368]: DEBUG nova.compute.manager [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Refreshing instance network info cache due to event network-changed-d229b53b-0e94-41c9-b8e9-0b0eaa366d3c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 939.939848] env[62368]: DEBUG oslo_concurrency.lockutils [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] Acquiring lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.991326] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d24e53f4-c0e6-4c30-aefa-c3e848c61e58 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "e427d283-1f8f-4c7a-869c-4cd48a595470" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.822s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.032593] env[62368]: DEBUG nova.objects.base [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 940.032827] env[62368]: DEBUG nova.network.neutron [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 940.096460] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52da14bc-8389-f15b-ac31-006fae86077d, 'name': SearchDatastore_Task, 'duration_secs': 0.01151} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.096783] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.097063] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac/e6dd9825-f841-49ac-a75a-83d9de3858ac.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 940.097329] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-47943297-f630-40fa-953d-d83ffdd01b94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.101781] env[62368]: DEBUG nova.policy [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68c2209271e043b698fd3794955f40bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcecae6c09124e9fa4dcd5d63833b0bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.105234] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 940.105234] env[62368]: value = "task-1198839" [ 940.105234] env[62368]: _type = "Task" [ 940.105234] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.113844] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.135670] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.136035] env[62368]: DEBUG nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Instance network_info: |[{"id": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "address": "fa:16:3e:b0:23:0b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd229b53b-0e", "ovs_interfaceid": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.136469] env[62368]: DEBUG oslo_concurrency.lockutils [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] Acquired lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.136544] env[62368]: DEBUG nova.network.neutron [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Refreshing network info cache for port d229b53b-0e94-41c9-b8e9-0b0eaa366d3c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 940.137792] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:23:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd229b53b-0e94-41c9-b8e9-0b0eaa366d3c', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.147103] env[62368]: DEBUG oslo.service.loopingcall [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.150583] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 940.151086] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a745625-2a1f-4aeb-965e-5cdad9a5be63 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.172548] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.172548] env[62368]: value = "task-1198840" [ 940.172548] env[62368]: _type = "Task" [ 940.172548] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.181541] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198840, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.330324] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.330483] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.342253] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c3262a6b-c12c-4400-bc73-e1a1dccaa3a8 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 32.524s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.342311] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 8.705s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.342617] env[62368]: INFO nova.compute.manager [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Unshelving [ 940.368070] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a5f6d1da-399f-4e19-af44-54b0f331afe2 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "5af32858-dc9a-4380-827a-daf384c29bed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.077s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.423707] env[62368]: DEBUG nova.network.neutron [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updated VIF entry in instance network info cache for port d229b53b-0e94-41c9-b8e9-0b0eaa366d3c. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 940.424118] env[62368]: DEBUG nova.network.neutron [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance_info_cache with network_info: [{"id": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "address": "fa:16:3e:b0:23:0b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd229b53b-0e", "ovs_interfaceid": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.616968] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51142} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.617703] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac/e6dd9825-f841-49ac-a75a-83d9de3858ac.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 940.618148] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.618541] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25bcea3d-30c8-424e-ba16-f735af20cd45 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.625402] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 940.625402] env[62368]: value = "task-1198841" [ 940.625402] env[62368]: _type = "Task" [ 940.625402] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.634818] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198841, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.682770] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198840, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.833672] env[62368]: DEBUG nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 940.928485] env[62368]: DEBUG oslo_concurrency.lockutils [req-c16cdaaa-a6d2-4db8-8dec-2bca62b19daf req-4dfb7f7a-5e1b-4c3a-b6ce-e1e2b66b12eb service nova] Releasing lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.107418] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8cb5f8-5973-4817-aba5-97458cf08513 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.115862] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b4f30c-dd7f-4554-b0d2-ccdd569c4896 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.149632] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2b53d1-1ca2-4cb5-8def-6d27a1cbe5f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.171879] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d702e030-547b-478e-ba5c-191a9f8baa1b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.176272] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198841, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076478} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.176553] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.180915] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b697688d-fdd1-4721-9da0-9617d1855d9c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.191721] env[62368]: DEBUG nova.compute.provider_tree [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.198836] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198840, 'name': CreateVM_Task, 'duration_secs': 0.660632} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.208013] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 941.216825] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac/e6dd9825-f841-49ac-a75a-83d9de3858ac.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.218176] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.218349] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.218678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.218917] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94021391-1b20-4730-a972-42caccfda966 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.233115] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac690e5a-d6b0-4ac9-9270-80051dd217b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.239008] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 941.239008] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529f0963-322e-fd50-aa09-827dd208808a" [ 941.239008] env[62368]: _type = "Task" [ 941.239008] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.244254] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 941.244254] env[62368]: value = "task-1198842" [ 941.244254] env[62368]: _type = "Task" [ 941.244254] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.250746] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529f0963-322e-fd50-aa09-827dd208808a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.257071] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198842, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.357224] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.368290] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.686099] env[62368]: DEBUG nova.compute.manager [req-8952d707-255e-419f-9512-f0b99d8425d1 req-9478eeb0-5632-457e-aea1-bc181f7c620c service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-vif-plugged-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.686404] env[62368]: DEBUG oslo_concurrency.lockutils [req-8952d707-255e-419f-9512-f0b99d8425d1 req-9478eeb0-5632-457e-aea1-bc181f7c620c service nova] Acquiring lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.686555] env[62368]: DEBUG oslo_concurrency.lockutils [req-8952d707-255e-419f-9512-f0b99d8425d1 req-9478eeb0-5632-457e-aea1-bc181f7c620c service nova] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.686735] env[62368]: DEBUG oslo_concurrency.lockutils [req-8952d707-255e-419f-9512-f0b99d8425d1 req-9478eeb0-5632-457e-aea1-bc181f7c620c service nova] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.686922] env[62368]: DEBUG nova.compute.manager [req-8952d707-255e-419f-9512-f0b99d8425d1 req-9478eeb0-5632-457e-aea1-bc181f7c620c service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] No waiting events found dispatching network-vif-plugged-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.687153] env[62368]: WARNING nova.compute.manager [req-8952d707-255e-419f-9512-f0b99d8425d1 req-9478eeb0-5632-457e-aea1-bc181f7c620c service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received unexpected event network-vif-plugged-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f for instance with vm_state active and task_state None. [ 941.700430] env[62368]: DEBUG nova.scheduler.client.report [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.749175] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529f0963-322e-fd50-aa09-827dd208808a, 'name': SearchDatastore_Task, 'duration_secs': 0.016509} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.752320] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.752570] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.752804] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.752957] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.753550] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.754043] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e7b6321-3d65-454b-95bd-4760f2bea4f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.758806] env[62368]: DEBUG nova.network.neutron [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Successfully updated port: d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.763238] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198842, 'name': ReconfigVM_Task, 'duration_secs': 0.511992} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.764471] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfigured VM instance instance-0000005c to attach disk [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac/e6dd9825-f841-49ac-a75a-83d9de3858ac.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.765141] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.765330] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 941.766271] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bc32de3-92c7-4587-8f50-796be8b9d63f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.768022] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb6a2948-9834-4e92-8c5b-8c4e606eb08e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.774572] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 941.774572] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5246cc35-a73d-076e-1d11-88ac0884645c" [ 941.774572] env[62368]: _type = "Task" [ 941.774572] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.776958] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 941.776958] env[62368]: value = "task-1198843" [ 941.776958] env[62368]: _type = "Task" [ 941.776958] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.793264] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198843, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.793610] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5246cc35-a73d-076e-1d11-88ac0884645c, 'name': SearchDatastore_Task, 'duration_secs': 0.013371} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.794809] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fde4c8c-46f4-413b-a741-17e3f9b19cf7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.800531] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 941.800531] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f48ec4-5652-f5cc-5f30-bdad7acc21b8" [ 941.800531] env[62368]: _type = "Task" [ 941.800531] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.811326] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f48ec4-5652-f5cc-5f30-bdad7acc21b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.205909] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.206569] env[62368]: DEBUG nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.209659] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.142s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.211324] env[62368]: INFO nova.compute.claims [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.267494] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.268462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.268462] env[62368]: DEBUG nova.network.neutron [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 942.291317] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198843, 'name': Rename_Task, 'duration_secs': 0.224949} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.291596] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 942.291843] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee711a67-9d6d-4069-9e01-4c6d7299d271 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.299886] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 942.299886] env[62368]: value = "task-1198844" [ 942.299886] env[62368]: _type = "Task" [ 942.299886] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.318885] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198844, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.319160] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f48ec4-5652-f5cc-5f30-bdad7acc21b8, 'name': SearchDatastore_Task, 'duration_secs': 0.016697} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.319426] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.319674] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2/72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 942.319945] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-932f9c4d-740a-48be-a49b-74fd475a6fb0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.327138] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 942.327138] env[62368]: value = "task-1198845" [ 942.327138] env[62368]: _type = "Task" [ 942.327138] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.337534] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.716698] env[62368]: DEBUG nova.compute.utils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.720887] env[62368]: DEBUG nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.721118] env[62368]: DEBUG nova.network.neutron [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 942.776942] env[62368]: DEBUG nova.policy [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c6e1b84100c4d4bb47f5d95272b62b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b7091fbcb2f4879b78075ed06b95c0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.813115] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198844, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.814212] env[62368]: WARNING nova.network.neutron [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] 3ee164de-30af-4404-935a-a7f988d824b1 already exists in list: networks containing: ['3ee164de-30af-4404-935a-a7f988d824b1']. ignoring it [ 942.839684] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198845, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.125345] env[62368]: DEBUG nova.network.neutron [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1b97d93-2fbd-4ff2-948f-6f197d6ed77f", "address": "fa:16:3e:26:cf:d1", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1b97d93-2f", "ovs_interfaceid": "d1b97d93-2fbd-4ff2-948f-6f197d6ed77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.221689] env[62368]: DEBUG nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.233921] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Acquiring lock "c1b238e7-3636-47db-b86e-3fcfe917f705" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.234271] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "c1b238e7-3636-47db-b86e-3fcfe917f705" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.257863] env[62368]: DEBUG nova.network.neutron [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Successfully created port: 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.314351] env[62368]: DEBUG oslo_vmware.api [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198844, 'name': PowerOnVM_Task, 'duration_secs': 0.847143} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.314626] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 943.314824] env[62368]: INFO nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Took 8.02 seconds to spawn the instance on the hypervisor. [ 943.315016] env[62368]: DEBUG nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.315829] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d0d90e-65bd-4b38-a19c-4abb9d54ad85 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.338774] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198845, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619361} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.339044] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2/72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 943.339267] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.339518] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07f27e6b-907a-4b89-9f2c-28cd9da7a7dd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.347674] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 943.347674] env[62368]: value = "task-1198846" [ 943.347674] env[62368]: _type = "Task" [ 943.347674] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.362838] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.468705] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349167a4-8510-406e-80fc-12dff7dc2e5a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.477171] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f834473-6dd3-4347-a1f6-b8941aa795e0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.510642] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd06ce65-e29c-4c47-bf7e-68da0eed8856 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.518913] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1a08c8-8a38-4fcd-ad31-5f25cc0004e1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.532963] env[62368]: DEBUG nova.compute.provider_tree [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.628234] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.629103] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.629328] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.630457] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42ec507-ebd1-4167-9098-11c9b220df2d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.650465] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.650744] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.650853] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.651049] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.651202] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.651352] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.651561] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.651719] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.651883] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.652063] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.652244] env[62368]: DEBUG nova.virt.hardware [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.659961] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Reconfiguring VM to attach interface {{(pid=62368) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 943.659961] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e05ccfb-87f3-4e4b-b27d-3c5d2cebc01c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.676059] env[62368]: DEBUG oslo_vmware.api [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 943.676059] env[62368]: value = "task-1198847" [ 943.676059] env[62368]: _type = "Task" [ 943.676059] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.683654] env[62368]: DEBUG oslo_vmware.api [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198847, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.719634] env[62368]: DEBUG nova.compute.manager [req-eab7bbaa-61e3-4963-ac86-344bd60bcfe9 req-43f4096e-6d8b-42d9-8b01-245d8ffb4c1d service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-changed-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.719899] env[62368]: DEBUG nova.compute.manager [req-eab7bbaa-61e3-4963-ac86-344bd60bcfe9 req-43f4096e-6d8b-42d9-8b01-245d8ffb4c1d service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing instance network info cache due to event network-changed-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.720180] env[62368]: DEBUG oslo_concurrency.lockutils [req-eab7bbaa-61e3-4963-ac86-344bd60bcfe9 req-43f4096e-6d8b-42d9-8b01-245d8ffb4c1d service nova] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.720180] env[62368]: DEBUG oslo_concurrency.lockutils [req-eab7bbaa-61e3-4963-ac86-344bd60bcfe9 req-43f4096e-6d8b-42d9-8b01-245d8ffb4c1d service nova] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.720388] env[62368]: DEBUG nova.network.neutron [req-eab7bbaa-61e3-4963-ac86-344bd60bcfe9 req-43f4096e-6d8b-42d9-8b01-245d8ffb4c1d service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing network info cache for port d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 943.736636] env[62368]: DEBUG nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 943.834799] env[62368]: INFO nova.compute.manager [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Took 15.62 seconds to build instance. [ 943.860320] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072415} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.860650] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.861481] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780a5312-75bb-440c-a4c1-bbda5b375642 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.886903] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2/72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.887390] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-113d93d6-1ee7-40a9-8022-6b41c2ea11a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.910037] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 943.910037] env[62368]: value = "task-1198848" [ 943.910037] env[62368]: _type = "Task" [ 943.910037] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.920855] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198848, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.035836] env[62368]: DEBUG nova.scheduler.client.report [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.190087] env[62368]: DEBUG oslo_vmware.api [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.234881] env[62368]: DEBUG nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.261442] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.267225] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.267394] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.267482] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.267682] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.267833] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.267985] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.268211] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.268429] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.268643] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.268826] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.269039] env[62368]: DEBUG nova.virt.hardware [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.269885] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25b13fd-6539-47ce-b165-6a024ab0e5ff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.279235] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3399b0a8-bd06-4cab-bb8e-b494979ec568 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.338692] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b0245fa6-5ee7-44d1-b606-2cff225f193f tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.135s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.419793] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198848, 'name': ReconfigVM_Task, 'duration_secs': 0.331095} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.420144] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2/72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.420804] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ae333b1-a760-4047-89b4-3b321ae2bda8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.428935] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 944.428935] env[62368]: value = "task-1198849" [ 944.428935] env[62368]: _type = "Task" [ 944.428935] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.433082] env[62368]: DEBUG nova.network.neutron [req-eab7bbaa-61e3-4963-ac86-344bd60bcfe9 req-43f4096e-6d8b-42d9-8b01-245d8ffb4c1d service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updated VIF entry in instance network info cache for port d1b97d93-2fbd-4ff2-948f-6f197d6ed77f. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 944.433512] env[62368]: DEBUG nova.network.neutron [req-eab7bbaa-61e3-4963-ac86-344bd60bcfe9 req-43f4096e-6d8b-42d9-8b01-245d8ffb4c1d service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1b97d93-2fbd-4ff2-948f-6f197d6ed77f", "address": "fa:16:3e:26:cf:d1", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1b97d93-2f", "ovs_interfaceid": "d1b97d93-2fbd-4ff2-948f-6f197d6ed77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.440623] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198849, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.541503] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.542187] env[62368]: DEBUG nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 944.545107] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.218s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.545378] env[62368]: DEBUG nova.objects.instance [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'resources' on Instance uuid 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.687985] env[62368]: DEBUG oslo_vmware.api [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198847, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.936780] env[62368]: DEBUG oslo_concurrency.lockutils [req-eab7bbaa-61e3-4963-ac86-344bd60bcfe9 req-43f4096e-6d8b-42d9-8b01-245d8ffb4c1d service nova] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.940610] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198849, 'name': Rename_Task, 'duration_secs': 0.151956} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.940958] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 944.941279] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba67a9fa-6801-4c38-98bd-5547a75ee25e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.949246] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 944.949246] env[62368]: value = "task-1198850" [ 944.949246] env[62368]: _type = "Task" [ 944.949246] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.964510] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.048462] env[62368]: DEBUG nova.compute.utils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.050120] env[62368]: DEBUG nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.050313] env[62368]: DEBUG nova.network.neutron [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 945.052590] env[62368]: DEBUG nova.objects.instance [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'numa_topology' on Instance uuid 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.096669] env[62368]: DEBUG nova.policy [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87c366f937e2485c8b0dc317d6acb8f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '430e72f3b08a4207a2fb4465ff160714', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 945.190184] env[62368]: DEBUG oslo_vmware.api [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198847, 'name': ReconfigVM_Task, 'duration_secs': 1.045057} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.193058] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.193058] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Reconfigured VM to attach interface {{(pid=62368) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 945.408584] env[62368]: DEBUG nova.network.neutron [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Successfully created port: d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.428710] env[62368]: DEBUG nova.network.neutron [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Successfully updated port: 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.465069] env[62368]: DEBUG oslo_vmware.api [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198850, 'name': PowerOnVM_Task, 'duration_secs': 0.453536} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.465069] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 945.465069] env[62368]: INFO nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Took 7.69 seconds to spawn the instance on the hypervisor. [ 945.465069] env[62368]: DEBUG nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.465069] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc3561e-1280-4b10-8814-99a0f4a6c088 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.558035] env[62368]: DEBUG nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 945.565024] env[62368]: DEBUG nova.objects.base [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Object Instance<3209ccfd-34e7-4c95-ba24-3d86fcdbae6e> lazy-loaded attributes: resources,numa_topology {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 945.697470] env[62368]: DEBUG oslo_concurrency.lockutils [None req-752975de-e74b-47ca-9cae-5dc3642ae9c4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-a5cbadbd-20dd-4514-8867-20243af5db0c-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.761s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.839479] env[62368]: DEBUG nova.compute.manager [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received event network-vif-plugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.839479] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.839841] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.839841] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.840133] env[62368]: DEBUG nova.compute.manager [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] No waiting events found dispatching network-vif-plugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 945.841286] env[62368]: WARNING nova.compute.manager [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received unexpected event network-vif-plugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 for instance with vm_state building and task_state spawning. [ 945.841286] env[62368]: DEBUG nova.compute.manager [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received event network-changed-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.841286] env[62368]: DEBUG nova.compute.manager [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Refreshing instance network info cache due to event network-changed-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 945.841286] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Acquiring lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.841496] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Acquired lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.841496] env[62368]: DEBUG nova.network.neutron [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Refreshing network info cache for port 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.849170] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107035bf-8fd4-4332-9e07-5bfce205087d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.858773] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf34dde1-ea6a-4a0f-8299-42e4c1869176 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.891877] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f7c477-66bd-497b-a827-15cbde221687 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.900458] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189a37e1-4d94-4087-b2c6-d2fd05acd182 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.916391] env[62368]: DEBUG nova.compute.provider_tree [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.932230] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.985043] env[62368]: INFO nova.compute.manager [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Took 16.94 seconds to build instance. [ 946.382121] env[62368]: DEBUG nova.network.neutron [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 946.420402] env[62368]: DEBUG nova.scheduler.client.report [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.488122] env[62368]: DEBUG oslo_concurrency.lockutils [None req-84e523f9-8cb5-4fd5-908b-514084b86291 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.462s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.528362] env[62368]: DEBUG nova.network.neutron [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.574908] env[62368]: DEBUG nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 946.602612] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.602905] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.603111] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.603346] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.603567] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.603755] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.604025] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.604224] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.606142] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.606142] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.606142] env[62368]: DEBUG nova.virt.hardware [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.606142] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc04e3a3-c8f9-47a2-bc25-6484c60ff6f8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.618238] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e86bab2-4147-4110-9801-f38725b5478b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.927644] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.380s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.928179] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.571s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.929787] env[62368]: INFO nova.compute.claims [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.032409] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Releasing lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.032828] env[62368]: DEBUG nova.compute.manager [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Received event network-changed-05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.033110] env[62368]: DEBUG nova.compute.manager [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Refreshing instance network info cache due to event network-changed-05119e53-57b5-4b7d-a056-3bb898352179. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.033521] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Acquiring lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.033781] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Acquired lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.034077] env[62368]: DEBUG nova.network.neutron [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Refreshing network info cache for port 05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 947.035636] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.035858] env[62368]: DEBUG nova.network.neutron [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.069726] env[62368]: DEBUG nova.compute.manager [req-c8bc4ece-1e6d-4b7a-b973-a28323ed586b req-a2811f92-7f1e-47ec-af03-0a822c071e2b service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Received event network-vif-plugged-d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.070035] env[62368]: DEBUG oslo_concurrency.lockutils [req-c8bc4ece-1e6d-4b7a-b973-a28323ed586b req-a2811f92-7f1e-47ec-af03-0a822c071e2b service nova] Acquiring lock "1d5a93f0-1068-4514-b01c-011e95f48498-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.070278] env[62368]: DEBUG oslo_concurrency.lockutils [req-c8bc4ece-1e6d-4b7a-b973-a28323ed586b req-a2811f92-7f1e-47ec-af03-0a822c071e2b service nova] Lock "1d5a93f0-1068-4514-b01c-011e95f48498-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.070474] env[62368]: DEBUG oslo_concurrency.lockutils [req-c8bc4ece-1e6d-4b7a-b973-a28323ed586b req-a2811f92-7f1e-47ec-af03-0a822c071e2b service nova] Lock "1d5a93f0-1068-4514-b01c-011e95f48498-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.070899] env[62368]: DEBUG nova.compute.manager [req-c8bc4ece-1e6d-4b7a-b973-a28323ed586b req-a2811f92-7f1e-47ec-af03-0a822c071e2b service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] No waiting events found dispatching network-vif-plugged-d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.071117] env[62368]: WARNING nova.compute.manager [req-c8bc4ece-1e6d-4b7a-b973-a28323ed586b req-a2811f92-7f1e-47ec-af03-0a822c071e2b service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Received unexpected event network-vif-plugged-d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 for instance with vm_state building and task_state spawning. [ 947.147374] env[62368]: DEBUG nova.network.neutron [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Successfully updated port: d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.212968] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "interface-a5cbadbd-20dd-4514-8867-20243af5db0c-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.213262] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-a5cbadbd-20dd-4514-8867-20243af5db0c-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.297789] env[62368]: DEBUG nova.compute.manager [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Stashing vm_state: active {{(pid=62368) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 947.439202] env[62368]: DEBUG oslo_concurrency.lockutils [None req-102d936e-0572-44a5-b719-6552f18380c7 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.679s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.440045] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 10.844s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.440273] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.440477] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.440646] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.443343] env[62368]: INFO nova.compute.manager [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Terminating instance [ 947.445022] env[62368]: DEBUG nova.compute.manager [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.445319] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 947.445537] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eda6a2ba-2176-4301-be73-349c2d1b332a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.456159] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4760cdb2-d02c-4359-a13b-4497da5c2a9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.488849] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e could not be found. [ 947.489077] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 947.489266] env[62368]: INFO nova.compute.manager [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 947.489512] env[62368]: DEBUG oslo.service.loopingcall [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.489824] env[62368]: DEBUG nova.compute.manager [-] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 947.489959] env[62368]: DEBUG nova.network.neutron [-] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 947.588204] env[62368]: DEBUG nova.network.neutron [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 947.650852] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.651143] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.651143] env[62368]: DEBUG nova.network.neutron [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.715912] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.716181] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.717097] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad92e34-d267-436f-8bc7-98339b2a2def {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.736825] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4977c68b-f393-4d96-bb44-9a002b4c0ea9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.766524] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Reconfiguring VM to detach interface {{(pid=62368) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 947.771840] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6225067a-139b-4646-ba13-1e8811fed245 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.801065] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 947.801065] env[62368]: value = "task-1198851" [ 947.801065] env[62368]: _type = "Task" [ 947.801065] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.816686] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.818115] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.900173] env[62368]: DEBUG nova.network.neutron [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [{"id": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "address": "fa:16:3e:a8:8e:88", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15eed45d-e0", "ovs_interfaceid": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.947041] env[62368]: DEBUG nova.network.neutron [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updated VIF entry in instance network info cache for port 05119e53-57b5-4b7d-a056-3bb898352179. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 947.947450] env[62368]: DEBUG nova.network.neutron [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [{"id": "05119e53-57b5-4b7d-a056-3bb898352179", "address": "fa:16:3e:3b:b9:de", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05119e53-57", "ovs_interfaceid": "05119e53-57b5-4b7d-a056-3bb898352179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.147138] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4b3cda-08de-464b-8a79-0f28aff7c897 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.157954] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fa79cd-9970-4f2a-a136-4918bbc44080 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.193042] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea80ca7-f301-4841-a21f-4bec94ff159c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.196359] env[62368]: DEBUG nova.network.neutron [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 948.203437] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11add91a-39d7-4699-864a-04e44f531cae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.217636] env[62368]: DEBUG nova.compute.provider_tree [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.294679] env[62368]: DEBUG nova.network.neutron [-] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.317193] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.350231] env[62368]: DEBUG nova.network.neutron [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updating instance_info_cache with network_info: [{"id": "d47a86a3-3fd6-48d2-8a80-e6b7f201bd76", "address": "fa:16:3e:21:07:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd47a86a3-3f", "ovs_interfaceid": "d47a86a3-3fd6-48d2-8a80-e6b7f201bd76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.402685] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.403054] env[62368]: DEBUG nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Instance network_info: |[{"id": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "address": "fa:16:3e:a8:8e:88", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15eed45d-e0", "ovs_interfaceid": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 948.403608] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:8e:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15eed45d-e09b-4aac-b3e2-e3b77ed1efe2', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.411813] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating folder: Project (6b7091fbcb2f4879b78075ed06b95c0d). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 948.412157] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a5de2e4-d693-460f-a0e2-66e277b20bd0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.424242] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Created folder: Project (6b7091fbcb2f4879b78075ed06b95c0d) in parent group-v259706. [ 948.424454] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating folder: Instances. Parent ref: group-v259866. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 948.424704] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7237e45b-2d23-41ec-a1cb-811c7d9d2e46 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.435149] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Created folder: Instances in parent group-v259866. [ 948.435478] env[62368]: DEBUG oslo.service.loopingcall [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.435689] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 948.463478] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b1fb479-a15e-4446-ae53-0dd512a7ec10 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.463478] env[62368]: DEBUG oslo_concurrency.lockutils [req-f11bdb26-60b0-4262-ab95-74b5b67b124f req-07ad1e67-9fa6-4725-8500-1e57daa5d44e service nova] Releasing lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.463478] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.463478] env[62368]: value = "task-1198854" [ 948.463478] env[62368]: _type = "Task" [ 948.463478] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.470778] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198854, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.722419] env[62368]: DEBUG nova.scheduler.client.report [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.797659] env[62368]: INFO nova.compute.manager [-] [instance: 3209ccfd-34e7-4c95-ba24-3d86fcdbae6e] Took 1.31 seconds to deallocate network for instance. [ 948.815741] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.853104] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.853506] env[62368]: DEBUG nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Instance network_info: |[{"id": "d47a86a3-3fd6-48d2-8a80-e6b7f201bd76", "address": "fa:16:3e:21:07:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd47a86a3-3f", "ovs_interfaceid": "d47a86a3-3fd6-48d2-8a80-e6b7f201bd76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 948.853906] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:07:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd47a86a3-3fd6-48d2-8a80-e6b7f201bd76', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.861527] env[62368]: DEBUG oslo.service.loopingcall [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.861770] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 948.861997] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c3e5afc-eccf-489d-807f-40c90f756871 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.883450] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.883450] env[62368]: value = "task-1198855" [ 948.883450] env[62368]: _type = "Task" [ 948.883450] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.891464] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198855, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.968293] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198854, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.099039] env[62368]: DEBUG nova.compute.manager [req-9adfa89b-cb61-446c-841d-e27a48ad3d14 req-717ea66c-61f4-42b7-9d1d-9a47b26b1bd3 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Received event network-changed-d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.099287] env[62368]: DEBUG nova.compute.manager [req-9adfa89b-cb61-446c-841d-e27a48ad3d14 req-717ea66c-61f4-42b7-9d1d-9a47b26b1bd3 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Refreshing instance network info cache due to event network-changed-d47a86a3-3fd6-48d2-8a80-e6b7f201bd76. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.099506] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adfa89b-cb61-446c-841d-e27a48ad3d14 req-717ea66c-61f4-42b7-9d1d-9a47b26b1bd3 service nova] Acquiring lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.099675] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adfa89b-cb61-446c-841d-e27a48ad3d14 req-717ea66c-61f4-42b7-9d1d-9a47b26b1bd3 service nova] Acquired lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.099856] env[62368]: DEBUG nova.network.neutron [req-9adfa89b-cb61-446c-841d-e27a48ad3d14 req-717ea66c-61f4-42b7-9d1d-9a47b26b1bd3 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Refreshing network info cache for port d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 949.227648] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.228251] env[62368]: DEBUG nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 949.230721] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.863s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.230938] env[62368]: DEBUG nova.objects.instance [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lazy-loading 'pci_requests' on Instance uuid eea21546-fbbf-4440-829c-8583c4ccabb6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.316149] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.393448] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198855, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.468375] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198854, 'name': CreateVM_Task, 'duration_secs': 0.649458} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.468560] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 949.469304] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.469442] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.469811] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.470088] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2da89c1-e396-493a-9383-09abf81dbfef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.475074] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 949.475074] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ebf209-57be-61a9-ec2a-ec6d634cffc2" [ 949.475074] env[62368]: _type = "Task" [ 949.475074] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.482666] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ebf209-57be-61a9-ec2a-ec6d634cffc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.733723] env[62368]: DEBUG nova.compute.utils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.740435] env[62368]: DEBUG nova.objects.instance [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lazy-loading 'numa_topology' on Instance uuid eea21546-fbbf-4440-829c-8583c4ccabb6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.741547] env[62368]: DEBUG nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.741711] env[62368]: DEBUG nova.network.neutron [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 949.823818] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.824289] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9f34c227-1e8f-406c-83bc-a4c9b1b4bdb5 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "3209ccfd-34e7-4c95-ba24-3d86fcdbae6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.384s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.828552] env[62368]: DEBUG nova.policy [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c66c425264424e3eaa37c8058e62cb30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '026a3176abc34001ab3506445226777f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.894776] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198855, 'name': CreateVM_Task, 'duration_secs': 0.571486} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.894935] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 949.895739] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.953120] env[62368]: DEBUG nova.network.neutron [req-9adfa89b-cb61-446c-841d-e27a48ad3d14 req-717ea66c-61f4-42b7-9d1d-9a47b26b1bd3 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updated VIF entry in instance network info cache for port d47a86a3-3fd6-48d2-8a80-e6b7f201bd76. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 949.953591] env[62368]: DEBUG nova.network.neutron [req-9adfa89b-cb61-446c-841d-e27a48ad3d14 req-717ea66c-61f4-42b7-9d1d-9a47b26b1bd3 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updating instance_info_cache with network_info: [{"id": "d47a86a3-3fd6-48d2-8a80-e6b7f201bd76", "address": "fa:16:3e:21:07:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd47a86a3-3f", "ovs_interfaceid": "d47a86a3-3fd6-48d2-8a80-e6b7f201bd76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.986865] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ebf209-57be-61a9-ec2a-ec6d634cffc2, 'name': SearchDatastore_Task, 'duration_secs': 0.041645} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.987162] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.987419] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 949.987668] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.987820] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.988011] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 949.988311] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.988646] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.988880] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-066075e8-9c74-477c-80f3-7f07de55edc2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.991243] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39ef59a3-0940-4c67-ae70-36aae1511ffe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.996636] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 949.996636] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52dfffbe-3b69-b93f-1885-bdfaa82946f6" [ 949.996636] env[62368]: _type = "Task" [ 949.996636] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.001138] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.001331] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 950.002327] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83be3851-ed79-4aa5-bed9-d5c4d2d55cf8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.008041] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52dfffbe-3b69-b93f-1885-bdfaa82946f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.009355] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 950.009355] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5271b1db-55a2-bdab-c631-9dccf5e6cdc9" [ 950.009355] env[62368]: _type = "Task" [ 950.009355] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.016900] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5271b1db-55a2-bdab-c631-9dccf5e6cdc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.230939] env[62368]: DEBUG nova.network.neutron [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Successfully created port: f1a74431-d8b5-4107-9e82-02521ef5e35b {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.242968] env[62368]: DEBUG nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 950.246258] env[62368]: INFO nova.compute.claims [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.318331] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.456810] env[62368]: DEBUG oslo_concurrency.lockutils [req-9adfa89b-cb61-446c-841d-e27a48ad3d14 req-717ea66c-61f4-42b7-9d1d-9a47b26b1bd3 service nova] Releasing lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.512025] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52dfffbe-3b69-b93f-1885-bdfaa82946f6, 'name': SearchDatastore_Task, 'duration_secs': 0.020569} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.515530] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.515852] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.516104] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.522332] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5271b1db-55a2-bdab-c631-9dccf5e6cdc9, 'name': SearchDatastore_Task, 'duration_secs': 0.01088} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.523171] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a70d9e2f-ccc5-43d0-8bb3-ffbb946e96a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.528615] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 950.528615] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520d219f-72fd-0213-bc5c-1eb1cd89c2ea" [ 950.528615] env[62368]: _type = "Task" [ 950.528615] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.536933] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520d219f-72fd-0213-bc5c-1eb1cd89c2ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.817962] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.039620] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520d219f-72fd-0213-bc5c-1eb1cd89c2ea, 'name': SearchDatastore_Task, 'duration_secs': 0.009697} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.039898] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.040225] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] fa637a2e-a113-4d23-8924-728ba75f202c/fa637a2e-a113-4d23-8924-728ba75f202c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 951.040517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.040715] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.041052] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f1d7ea3-edb6-41ef-8f05-69bd0fc1dc94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.044896] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fa7a69f-e60f-4aa6-9337-84355bb9de38 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.054465] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 951.054465] env[62368]: value = "task-1198856" [ 951.054465] env[62368]: _type = "Task" [ 951.054465] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.067195] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198856, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.068680] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.068899] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 951.069649] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0f739a7-07a8-4d29-9d83-f0211f7e9d49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.076983] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 951.076983] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52d90cd7-ec9b-e2ac-1d69-77ca000eec16" [ 951.076983] env[62368]: _type = "Task" [ 951.076983] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.084868] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d90cd7-ec9b-e2ac-1d69-77ca000eec16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.153097] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "84bb7208-670f-4e1d-862a-a480f7db0f42" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.153469] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.257732] env[62368]: DEBUG nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 951.287592] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.287933] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.288194] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.288433] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.288647] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.288889] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.289217] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.289430] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.289655] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.289853] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.290116] env[62368]: DEBUG nova.virt.hardware [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.291508] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20a85f7-39f5-41ba-baba-26fae1c9477b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.302365] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039a2d43-1547-482d-aefd-cfec6bcc3cc5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.332280] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.534027] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d92ba7e-9863-4c4d-b396-c6625bfd81d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.542714] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7758e5-fed2-4fac-b626-cb255a547c2b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.577941] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3279baf-6faf-4103-9941-ae5fd33eb65b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.279650] env[62368]: DEBUG nova.network.neutron [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Successfully updated port: f1a74431-d8b5-4107-9e82-02521ef5e35b {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 952.283923] env[62368]: DEBUG nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 952.285029] env[62368]: DEBUG nova.compute.manager [req-c6336290-3dc6-44c2-9575-25de495be8fb req-0b78be6f-ec21-42da-8f41-535b58b29f76 service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Received event network-vif-plugged-f1a74431-d8b5-4107-9e82-02521ef5e35b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.285245] env[62368]: DEBUG oslo_concurrency.lockutils [req-c6336290-3dc6-44c2-9575-25de495be8fb req-0b78be6f-ec21-42da-8f41-535b58b29f76 service nova] Acquiring lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.285483] env[62368]: DEBUG oslo_concurrency.lockutils [req-c6336290-3dc6-44c2-9575-25de495be8fb req-0b78be6f-ec21-42da-8f41-535b58b29f76 service nova] Lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.285675] env[62368]: DEBUG oslo_concurrency.lockutils [req-c6336290-3dc6-44c2-9575-25de495be8fb req-0b78be6f-ec21-42da-8f41-535b58b29f76 service nova] Lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.285832] env[62368]: DEBUG nova.compute.manager [req-c6336290-3dc6-44c2-9575-25de495be8fb req-0b78be6f-ec21-42da-8f41-535b58b29f76 service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] No waiting events found dispatching network-vif-plugged-f1a74431-d8b5-4107-9e82-02521ef5e35b {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 952.286298] env[62368]: WARNING nova.compute.manager [req-c6336290-3dc6-44c2-9575-25de495be8fb req-0b78be6f-ec21-42da-8f41-535b58b29f76 service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Received unexpected event network-vif-plugged-f1a74431-d8b5-4107-9e82-02521ef5e35b for instance with vm_state building and task_state spawning. [ 952.295496] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198856, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52849} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.296400] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] fa637a2e-a113-4d23-8924-728ba75f202c/fa637a2e-a113-4d23-8924-728ba75f202c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 952.297093] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.298225] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70744a7b-2647-4820-8ea1-a0ffed06ebf6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.305721] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9634897d-b5e0-45a3-8932-c33904590179 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.308537] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52d90cd7-ec9b-e2ac-1d69-77ca000eec16, 'name': SearchDatastore_Task, 'duration_secs': 0.023218} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.312569] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e0290b1-89a7-4d6d-b02f-ff6ba89f5512 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.315880] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.324305] env[62368]: DEBUG nova.compute.provider_tree [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.329702] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 952.329702] env[62368]: value = "task-1198857" [ 952.329702] env[62368]: _type = "Task" [ 952.329702] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.332312] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 952.332312] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f0cd71-ea43-17e1-b3c7-93a5ab0f3a25" [ 952.332312] env[62368]: _type = "Task" [ 952.332312] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.339619] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.345544] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f0cd71-ea43-17e1-b3c7-93a5ab0f3a25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.791077] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "refresh_cache-cf3601d5-2956-4301-a7cc-bc2d33bf3790" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.791077] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "refresh_cache-cf3601d5-2956-4301-a7cc-bc2d33bf3790" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.791077] env[62368]: DEBUG nova.network.neutron [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.797623] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.810022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.829057] env[62368]: DEBUG nova.scheduler.client.report [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.843197] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198857, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068387} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.846978] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.847692] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f0cd71-ea43-17e1-b3c7-93a5ab0f3a25, 'name': SearchDatastore_Task, 'duration_secs': 0.014266} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.847969] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44d2bb1-9667-4134-9036-d8a2d5d8c236 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.850267] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.850526] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498/1d5a93f0-1068-4514-b01c-011e95f48498.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 952.850976] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d8f5fe2-60b2-4b3c-891d-9b87a13e5feb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.874606] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] fa637a2e-a113-4d23-8924-728ba75f202c/fa637a2e-a113-4d23-8924-728ba75f202c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.876389] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f530d358-755b-4973-b7ea-e2da853b2ae9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.890665] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 952.890665] env[62368]: value = "task-1198858" [ 952.890665] env[62368]: _type = "Task" [ 952.890665] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.898037] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 952.898037] env[62368]: value = "task-1198859" [ 952.898037] env[62368]: _type = "Task" [ 952.898037] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.900650] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198858, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.909284] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198859, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.300900] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.336356] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.105s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.339059] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.078s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.341114] env[62368]: INFO nova.compute.claims [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.345849] env[62368]: DEBUG nova.network.neutron [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 953.404653] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198858, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.405934] env[62368]: INFO nova.network.neutron [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating port b2ebe893-6ba3-40da-8def-fb6cf1ab442e with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 953.423069] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.612311] env[62368]: DEBUG nova.network.neutron [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Updating instance_info_cache with network_info: [{"id": "f1a74431-d8b5-4107-9e82-02521ef5e35b", "address": "fa:16:3e:86:86:d7", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1a74431-d8", "ovs_interfaceid": "f1a74431-d8b5-4107-9e82-02521ef5e35b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.787860] env[62368]: DEBUG nova.compute.manager [req-3554e960-0432-4ebd-bcb8-3b0ca69c5ddc req-69c2aefb-4053-4688-9210-7caf0144fdcb service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Received event network-changed-f1a74431-d8b5-4107-9e82-02521ef5e35b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.788088] env[62368]: DEBUG nova.compute.manager [req-3554e960-0432-4ebd-bcb8-3b0ca69c5ddc req-69c2aefb-4053-4688-9210-7caf0144fdcb service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Refreshing instance network info cache due to event network-changed-f1a74431-d8b5-4107-9e82-02521ef5e35b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 953.788284] env[62368]: DEBUG oslo_concurrency.lockutils [req-3554e960-0432-4ebd-bcb8-3b0ca69c5ddc req-69c2aefb-4053-4688-9210-7caf0144fdcb service nova] Acquiring lock "refresh_cache-cf3601d5-2956-4301-a7cc-bc2d33bf3790" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.797806] env[62368]: DEBUG oslo_vmware.api [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198851, 'name': ReconfigVM_Task, 'duration_secs': 5.82809} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.798049] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.798253] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Reconfigured VM to detach interface {{(pid=62368) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 953.903206] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198858, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688378} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.903490] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498/1d5a93f0-1068-4514-b01c-011e95f48498.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 953.903712] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.904024] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e948a244-72de-4ab2-80e1-a79edba7a4fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.914327] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198859, 'name': ReconfigVM_Task, 'duration_secs': 0.635933} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.915675] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Reconfigured VM instance instance-0000005e to attach disk [datastore2] fa637a2e-a113-4d23-8924-728ba75f202c/fa637a2e-a113-4d23-8924-728ba75f202c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.916396] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 953.916396] env[62368]: value = "task-1198860" [ 953.916396] env[62368]: _type = "Task" [ 953.916396] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.916602] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e93ec63-9b74-4dd5-8146-3237496bcf88 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.928148] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198860, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.929856] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 953.929856] env[62368]: value = "task-1198861" [ 953.929856] env[62368]: _type = "Task" [ 953.929856] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.938600] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198861, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.116022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "refresh_cache-cf3601d5-2956-4301-a7cc-bc2d33bf3790" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.116022] env[62368]: DEBUG nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Instance network_info: |[{"id": "f1a74431-d8b5-4107-9e82-02521ef5e35b", "address": "fa:16:3e:86:86:d7", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1a74431-d8", "ovs_interfaceid": "f1a74431-d8b5-4107-9e82-02521ef5e35b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 954.116022] env[62368]: DEBUG oslo_concurrency.lockutils [req-3554e960-0432-4ebd-bcb8-3b0ca69c5ddc req-69c2aefb-4053-4688-9210-7caf0144fdcb service nova] Acquired lock "refresh_cache-cf3601d5-2956-4301-a7cc-bc2d33bf3790" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.116022] env[62368]: DEBUG nova.network.neutron [req-3554e960-0432-4ebd-bcb8-3b0ca69c5ddc req-69c2aefb-4053-4688-9210-7caf0144fdcb service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Refreshing network info cache for port f1a74431-d8b5-4107-9e82-02521ef5e35b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 954.117280] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:86:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8f40f5c4-c146-449c-884d-6f884dcf2acf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1a74431-d8b5-4107-9e82-02521ef5e35b', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.124572] env[62368]: DEBUG oslo.service.loopingcall [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.125637] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 954.127261] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7b6c6b9-0b98-430a-a91f-314d6ea13fed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.146344] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.146344] env[62368]: value = "task-1198862" [ 954.146344] env[62368]: _type = "Task" [ 954.146344] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.166955] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198862, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.429452] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198860, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075283} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.431937] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.432924] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7260e8e-0f51-4fc9-a463-50c5bb3523e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.443372] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198861, 'name': Rename_Task, 'duration_secs': 0.173425} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.454437] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 954.463008] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498/1d5a93f0-1068-4514-b01c-011e95f48498.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.463505] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b34066b1-a3d9-4e04-99fd-40b7f3649399 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.465082] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7aca5c4-176f-4ef4-9ef5-cf5b28c93c07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.488849] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 954.488849] env[62368]: value = "task-1198864" [ 954.488849] env[62368]: _type = "Task" [ 954.488849] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.489866] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 954.489866] env[62368]: value = "task-1198863" [ 954.489866] env[62368]: _type = "Task" [ 954.489866] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.504794] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.507673] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198864, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.620337] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46844488-cc99-4e2d-8106-e60a8f7cc3bb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.631547] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a598a0-4be4-47b4-a2c6-af6f703ba395 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.669845] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f871187-1591-4b6f-801b-1fd1e22238ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.681807] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbf2ccb-0b5d-47ee-8ae7-146d4bd3c8fb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.685716] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198862, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.698803] env[62368]: DEBUG nova.compute.provider_tree [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.007853] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198863, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.012613] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198864, 'name': ReconfigVM_Task, 'duration_secs': 0.397528} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.013063] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498/1d5a93f0-1068-4514-b01c-011e95f48498.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.013941] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c7aab48-cc88-40c6-940e-58fa086896ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.018151] env[62368]: DEBUG nova.network.neutron [req-3554e960-0432-4ebd-bcb8-3b0ca69c5ddc req-69c2aefb-4053-4688-9210-7caf0144fdcb service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Updated VIF entry in instance network info cache for port f1a74431-d8b5-4107-9e82-02521ef5e35b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 955.018666] env[62368]: DEBUG nova.network.neutron [req-3554e960-0432-4ebd-bcb8-3b0ca69c5ddc req-69c2aefb-4053-4688-9210-7caf0144fdcb service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Updating instance_info_cache with network_info: [{"id": "f1a74431-d8b5-4107-9e82-02521ef5e35b", "address": "fa:16:3e:86:86:d7", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1a74431-d8", "ovs_interfaceid": "f1a74431-d8b5-4107-9e82-02521ef5e35b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.022679] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 955.022679] env[62368]: value = "task-1198865" [ 955.022679] env[62368]: _type = "Task" [ 955.022679] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.034811] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198865, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.176948] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.177305] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.177459] env[62368]: DEBUG nova.network.neutron [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 955.179076] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198862, 'name': CreateVM_Task, 'duration_secs': 0.697743} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.179076] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 955.180397] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.180833] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.181385] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.181830] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f320dfc2-3cf3-44dc-bd76-1c30a922a0c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.188306] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 955.188306] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e77302-d103-2467-bd7f-1de641e08c0c" [ 955.188306] env[62368]: _type = "Task" [ 955.188306] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.193503] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.193640] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.193910] env[62368]: DEBUG nova.network.neutron [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 955.205090] env[62368]: DEBUG nova.scheduler.client.report [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.209377] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e77302-d103-2467-bd7f-1de641e08c0c, 'name': SearchDatastore_Task, 'duration_secs': 0.010654} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.210268] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.210508] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.210737] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.210886] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.211175] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.211622] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb7cb4bb-839c-4476-a85b-a28e7a82dc9d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.223397] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.223643] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 955.225238] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f224d54-9b89-4026-af8e-9f67833b48f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.231563] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 955.231563] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527007be-7667-7e61-2c58-25834d231cd5" [ 955.231563] env[62368]: _type = "Task" [ 955.231563] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.241537] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527007be-7667-7e61-2c58-25834d231cd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.506205] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198863, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.522225] env[62368]: DEBUG oslo_concurrency.lockutils [req-3554e960-0432-4ebd-bcb8-3b0ca69c5ddc req-69c2aefb-4053-4688-9210-7caf0144fdcb service nova] Releasing lock "refresh_cache-cf3601d5-2956-4301-a7cc-bc2d33bf3790" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.535318] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198865, 'name': Rename_Task, 'duration_secs': 0.163938} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.535669] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 955.535955] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82108bfc-eb2b-4623-b5e0-01a695cbb49d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.544359] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 955.544359] env[62368]: value = "task-1198866" [ 955.544359] env[62368]: _type = "Task" [ 955.544359] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.556102] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198866, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.712754] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.713148] env[62368]: DEBUG nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 955.716076] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 7.898s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.746353] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527007be-7667-7e61-2c58-25834d231cd5, 'name': SearchDatastore_Task, 'duration_secs': 0.009669} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.749812] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56446537-2798-4f69-998b-b67f753a63f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.758095] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 955.758095] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e4952d-ea04-a1c9-f7ab-f8179718f748" [ 955.758095] env[62368]: _type = "Task" [ 955.758095] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.767655] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e4952d-ea04-a1c9-f7ab-f8179718f748, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.814683] env[62368]: DEBUG nova.compute.manager [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received event network-vif-plugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.814896] env[62368]: DEBUG oslo_concurrency.lockutils [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] Acquiring lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.815122] env[62368]: DEBUG oslo_concurrency.lockutils [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.815296] env[62368]: DEBUG oslo_concurrency.lockutils [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.815503] env[62368]: DEBUG nova.compute.manager [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] No waiting events found dispatching network-vif-plugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.815694] env[62368]: WARNING nova.compute.manager [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received unexpected event network-vif-plugged-b2ebe893-6ba3-40da-8def-fb6cf1ab442e for instance with vm_state shelved_offloaded and task_state spawning. [ 955.815859] env[62368]: DEBUG nova.compute.manager [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received event network-changed-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.816044] env[62368]: DEBUG nova.compute.manager [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Refreshing instance network info cache due to event network-changed-b2ebe893-6ba3-40da-8def-fb6cf1ab442e. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.816297] env[62368]: DEBUG oslo_concurrency.lockutils [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] Acquiring lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.006378] env[62368]: DEBUG oslo_vmware.api [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198863, 'name': PowerOnVM_Task, 'duration_secs': 1.085847} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.007022] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 956.007022] env[62368]: INFO nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Took 11.77 seconds to spawn the instance on the hypervisor. [ 956.008867] env[62368]: DEBUG nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.008867] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04579e73-b50a-4b28-b6f4-0df823e0d710 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.025630] env[62368]: DEBUG nova.compute.manager [req-6b1de980-310d-4923-8a70-d8417758012a req-cec238d3-7abf-47fe-b03f-b2e01ddbe726 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.025630] env[62368]: DEBUG nova.compute.manager [req-6b1de980-310d-4923-8a70-d8417758012a req-cec238d3-7abf-47fe-b03f-b2e01ddbe726 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing instance network info cache due to event network-changed-c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.025630] env[62368]: DEBUG oslo_concurrency.lockutils [req-6b1de980-310d-4923-8a70-d8417758012a req-cec238d3-7abf-47fe-b03f-b2e01ddbe726 service nova] Acquiring lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.054013] env[62368]: DEBUG oslo_vmware.api [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198866, 'name': PowerOnVM_Task, 'duration_secs': 0.46459} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.054760] env[62368]: INFO nova.network.neutron [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Port d1b97d93-2fbd-4ff2-948f-6f197d6ed77f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 956.055623] env[62368]: DEBUG nova.network.neutron [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.056700] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 956.056927] env[62368]: INFO nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Took 9.48 seconds to spawn the instance on the hypervisor. [ 956.057165] env[62368]: DEBUG nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.059166] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a69c769-e79c-4d04-92b7-1a0ac84d303f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.090717] env[62368]: DEBUG nova.network.neutron [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating instance_info_cache with network_info: [{"id": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "address": "fa:16:3e:d1:c4:2e", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ebe893-6b", "ovs_interfaceid": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.220827] env[62368]: INFO nova.compute.claims [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.224986] env[62368]: DEBUG nova.compute.utils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.227025] env[62368]: DEBUG nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.227105] env[62368]: DEBUG nova.network.neutron [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 956.266712] env[62368]: DEBUG nova.policy [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e58f21fa5634a13946bd09e6997cb10', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46f27452524540b596fa32b9999dc208', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 956.272871] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e4952d-ea04-a1c9-f7ab-f8179718f748, 'name': SearchDatastore_Task, 'duration_secs': 0.012327} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.273185] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.273498] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] cf3601d5-2956-4301-a7cc-bc2d33bf3790/cf3601d5-2956-4301-a7cc-bc2d33bf3790.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 956.273770] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a72e9005-3fc8-4f92-b945-e864c324510a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.282437] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 956.282437] env[62368]: value = "task-1198867" [ 956.282437] env[62368]: _type = "Task" [ 956.282437] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.291752] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198867, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.532052] env[62368]: INFO nova.compute.manager [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Took 23.66 seconds to build instance. [ 956.558360] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.560551] env[62368]: DEBUG oslo_concurrency.lockutils [req-6b1de980-310d-4923-8a70-d8417758012a req-cec238d3-7abf-47fe-b03f-b2e01ddbe726 service nova] Acquired lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.560744] env[62368]: DEBUG nova.network.neutron [req-6b1de980-310d-4923-8a70-d8417758012a req-cec238d3-7abf-47fe-b03f-b2e01ddbe726 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Refreshing network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.579031] env[62368]: INFO nova.compute.manager [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Took 21.53 seconds to build instance. [ 956.593137] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.596402] env[62368]: DEBUG oslo_concurrency.lockutils [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] Acquired lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.596402] env[62368]: DEBUG nova.network.neutron [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Refreshing network info cache for port b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.625613] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='d6808e57ed5fb998680281bc193a1846',container_format='bare',created_at=2024-10-08T23:19:28Z,direct_url=,disk_format='vmdk',id=bc3d2926-26a5-47b1-b7eb-6a838c226d70,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1503100462-shelved',owner='e250606f3df240eb93ce45925b00ad43',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-08T23:19:44Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.625939] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.626420] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.626714] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.626954] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.627169] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.627475] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.627690] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.627920] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.628183] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.628393] env[62368]: DEBUG nova.virt.hardware [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.629437] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6482381f-bed9-4c24-94a1-200fff99a5d4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.639595] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7015346b-5a81-4c19-9c4b-ef6025b22c37 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.655276] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:c4:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7150f662-0cf1-44f9-ae14-d70f479649b6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2ebe893-6ba3-40da-8def-fb6cf1ab442e', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.663149] env[62368]: DEBUG oslo.service.loopingcall [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.663497] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 956.663741] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-420498a1-205b-4990-bd2b-6dc837706a6e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.685657] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.685657] env[62368]: value = "task-1198868" [ 956.685657] env[62368]: _type = "Task" [ 956.685657] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.695016] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198868, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.728053] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "interface-5633292c-9a74-4c2d-893b-6823f4a587a2-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.728053] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-5633292c-9a74-4c2d-893b-6823f4a587a2-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.728053] env[62368]: DEBUG nova.objects.instance [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'flavor' on Instance uuid 5633292c-9a74-4c2d-893b-6823f4a587a2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.728053] env[62368]: DEBUG nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 956.732926] env[62368]: INFO nova.compute.resource_tracker [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating resource usage from migration e3cd6a40-1f21-4511-adbd-66cdf0420c64 [ 956.792848] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198867, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.949460] env[62368]: DEBUG nova.network.neutron [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Successfully created port: d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 957.024118] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16af284-b1ae-4b12-9be3-71f061582d91 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.034202] env[62368]: DEBUG oslo_concurrency.lockutils [None req-162b1d09-af62-4a43-bffa-51cfb0e4abdc tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.170s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.035843] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c8ec1c-6394-4e2a-a4e6-7995f14cddb3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.067487] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e80d23e7-10f4-4966-a89b-a4284081cac4 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-a5cbadbd-20dd-4514-8867-20243af5db0c-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.854s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.071447] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7573688e-e26d-4266-8ab4-d2e4324fc403 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.080637] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1c25c4-ac38-4a65-ae8c-d1acf4f64dea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.085164] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6aff5de0-402c-4971-9420-15f860fbbd47 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.047s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.096317] env[62368]: DEBUG nova.compute.provider_tree [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.201545] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198868, 'name': CreateVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.237712] env[62368]: INFO nova.virt.block_device [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Booting with volume 0923c8dd-6948-46a8-a7ce-88b165401227 at /dev/sda [ 957.292441] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198867, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597151} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.294882] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] cf3601d5-2956-4301-a7cc-bc2d33bf3790/cf3601d5-2956-4301-a7cc-bc2d33bf3790.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 957.295835] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 957.296977] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64c3e4b0-2761-4daf-8136-85a72b6011f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.299132] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6000e003-3142-4e7f-ad89-860aa34dd5b7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.310098] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343e18bd-c892-4899-8a14-6f7f90f2fda6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.324012] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 957.324012] env[62368]: value = "task-1198869" [ 957.324012] env[62368]: _type = "Task" [ 957.324012] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.335312] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198869, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.336172] env[62368]: DEBUG nova.network.neutron [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updated VIF entry in instance network info cache for port b2ebe893-6ba3-40da-8def-fb6cf1ab442e. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 957.336542] env[62368]: DEBUG nova.network.neutron [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating instance_info_cache with network_info: [{"id": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "address": "fa:16:3e:d1:c4:2e", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ebe893-6b", "ovs_interfaceid": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.352165] env[62368]: DEBUG oslo_concurrency.lockutils [req-6f66fbb2-ee0d-4f61-8e34-1ad4df2c43bc req-9f583de1-4293-4015-bfe5-e62874dadee2 service nova] Releasing lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.352516] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-444a857e-074b-449d-bf87-fe1a88518c24 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.365743] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7897baad-7ecd-45d7-bd8e-e3ea887f1c10 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.402696] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb7c33b-4026-42df-a01d-387806c1b35b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.410829] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5781ec5a-c90b-4b51-beef-b88914a21906 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.426863] env[62368]: DEBUG nova.virt.block_device [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Updating existing volume attachment record: 5af1f9e6-f9c8-454c-8839-c012c8528734 {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 957.494594] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.496634] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.496634] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "58157ab1-80a4-427c-812b-f6fde1f8db68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.496634] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.496634] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.497737] env[62368]: INFO nova.compute.manager [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Terminating instance [ 957.503039] env[62368]: DEBUG nova.compute.manager [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.503039] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 957.503039] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20efeb2-87ca-4632-b6f0-8c2ecd1e89d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.510808] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 957.511092] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b303dc7-9646-4660-b7c4-a181a1be3212 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.518769] env[62368]: DEBUG oslo_vmware.api [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 957.518769] env[62368]: value = "task-1198870" [ 957.518769] env[62368]: _type = "Task" [ 957.518769] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.528920] env[62368]: DEBUG oslo_vmware.api [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.541773] env[62368]: DEBUG nova.objects.instance [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'pci_requests' on Instance uuid 5633292c-9a74-4c2d-893b-6823f4a587a2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.599526] env[62368]: DEBUG nova.network.neutron [req-6b1de980-310d-4923-8a70-d8417758012a req-cec238d3-7abf-47fe-b03f-b2e01ddbe726 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updated VIF entry in instance network info cache for port c06f647d-2b3f-4ee0-8221-d00fda253d7b. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 957.600054] env[62368]: DEBUG nova.network.neutron [req-6b1de980-310d-4923-8a70-d8417758012a req-cec238d3-7abf-47fe-b03f-b2e01ddbe726 service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [{"id": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "address": "fa:16:3e:6d:9b:cd", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc06f647d-2b", "ovs_interfaceid": "c06f647d-2b3f-4ee0-8221-d00fda253d7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.601833] env[62368]: DEBUG nova.scheduler.client.report [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.702537] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198868, 'name': CreateVM_Task, 'duration_secs': 0.610332} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.702760] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 957.703833] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.703958] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.704519] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.704883] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e55d886d-5a8b-4250-b5a0-88b26344fba5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.711470] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 957.711470] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fc0cfd-b251-a12e-3347-5d14431d4bcf" [ 957.711470] env[62368]: _type = "Task" [ 957.711470] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.723844] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fc0cfd-b251-a12e-3347-5d14431d4bcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.834576] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198869, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.31882} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.834893] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.836226] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8293ab0f-af33-4d52-8171-525292596834 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.859751] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] cf3601d5-2956-4301-a7cc-bc2d33bf3790/cf3601d5-2956-4301-a7cc-bc2d33bf3790.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.860061] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b4993b9-5a55-4912-a897-87fa7c39b35f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.882743] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 957.882743] env[62368]: value = "task-1198871" [ 957.882743] env[62368]: _type = "Task" [ 957.882743] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.892193] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198871, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.029593] env[62368]: DEBUG oslo_vmware.api [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198870, 'name': PowerOffVM_Task, 'duration_secs': 0.209419} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.029870] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 958.030055] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 958.030319] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d29520c7-d0e2-4612-80ab-6933b627ac6b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.044519] env[62368]: DEBUG nova.objects.base [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Object Instance<5633292c-9a74-4c2d-893b-6823f4a587a2> lazy-loaded attributes: flavor,pci_requests {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 958.044732] env[62368]: DEBUG nova.network.neutron [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 958.106338] env[62368]: DEBUG oslo_concurrency.lockutils [req-6b1de980-310d-4923-8a70-d8417758012a req-cec238d3-7abf-47fe-b03f-b2e01ddbe726 service nova] Releasing lock "refresh_cache-a5cbadbd-20dd-4514-8867-20243af5db0c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.107516] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.391s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.107654] env[62368]: INFO nova.compute.manager [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Migrating [ 958.116599] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.307s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.118126] env[62368]: INFO nova.compute.claims [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.178079] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 958.178505] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 958.178683] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Deleting the datastore file [datastore1] 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.179151] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dff4b563-67ad-4c5d-8309-09792008e43c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.191064] env[62368]: DEBUG oslo_vmware.api [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 958.191064] env[62368]: value = "task-1198873" [ 958.191064] env[62368]: _type = "Task" [ 958.191064] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.203958] env[62368]: DEBUG oslo_vmware.api [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198873, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.223800] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.224125] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Processing image bc3d2926-26a5-47b1-b7eb-6a838c226d70 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.224396] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70/bc3d2926-26a5-47b1-b7eb-6a838c226d70.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.224629] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70/bc3d2926-26a5-47b1-b7eb-6a838c226d70.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.224834] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.225179] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ccbc032-befd-4376-8a3a-b39da6bb935c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.247857] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.248190] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 958.249464] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f47389e-4356-4367-8c15-83c24cfe427e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.254549] env[62368]: DEBUG nova.policy [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68c2209271e043b698fd3794955f40bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcecae6c09124e9fa4dcd5d63833b0bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.261425] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 958.261425] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5218a1e5-daf1-0615-d745-c26e85b5ce35" [ 958.261425] env[62368]: _type = "Task" [ 958.261425] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.272121] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5218a1e5-daf1-0615-d745-c26e85b5ce35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.396148] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.629918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.630302] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.630552] env[62368]: DEBUG nova.network.neutron [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 958.705182] env[62368]: DEBUG oslo_vmware.api [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198873, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.303654} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.705836] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.706154] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 958.706456] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 958.706748] env[62368]: INFO nova.compute.manager [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Took 1.21 seconds to destroy the instance on the hypervisor. [ 958.707144] env[62368]: DEBUG oslo.service.loopingcall [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.707458] env[62368]: DEBUG nova.compute.manager [-] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 958.707610] env[62368]: DEBUG nova.network.neutron [-] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 958.774342] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Preparing fetch location {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 958.774681] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Fetch image to [datastore1] OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf/OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf.vmdk {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 958.774863] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Downloading stream optimized image bc3d2926-26a5-47b1-b7eb-6a838c226d70 to [datastore1] OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf/OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf.vmdk on the data store datastore1 as vApp {{(pid=62368) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 958.775045] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Downloading image file data bc3d2926-26a5-47b1-b7eb-6a838c226d70 to the ESX as VM named 'OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf' {{(pid=62368) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 958.871920] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 958.871920] env[62368]: value = "resgroup-9" [ 958.871920] env[62368]: _type = "ResourcePool" [ 958.871920] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 958.872355] env[62368]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7ece346f-2df5-47cf-8fe5-4a1201881686 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.900335] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198871, 'name': ReconfigVM_Task, 'duration_secs': 0.648412} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.901811] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Reconfigured VM instance instance-00000060 to attach disk [datastore2] cf3601d5-2956-4301-a7cc-bc2d33bf3790/cf3601d5-2956-4301-a7cc-bc2d33bf3790.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.902612] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lease: (returnval){ [ 958.902612] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e99a37-54c4-3caa-6d86-f7405ce52677" [ 958.902612] env[62368]: _type = "HttpNfcLease" [ 958.902612] env[62368]: } obtained for vApp import into resource pool (val){ [ 958.902612] env[62368]: value = "resgroup-9" [ 958.902612] env[62368]: _type = "ResourcePool" [ 958.902612] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 958.902853] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the lease: (returnval){ [ 958.902853] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e99a37-54c4-3caa-6d86-f7405ce52677" [ 958.902853] env[62368]: _type = "HttpNfcLease" [ 958.902853] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 958.903034] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6b47083-273e-4096-8392-0bd9c411b980 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.918033] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.918033] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e99a37-54c4-3caa-6d86-f7405ce52677" [ 958.918033] env[62368]: _type = "HttpNfcLease" [ 958.918033] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.918033] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 958.918033] env[62368]: value = "task-1198875" [ 958.918033] env[62368]: _type = "Task" [ 958.918033] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.926415] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198875, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.011784] env[62368]: DEBUG nova.compute.manager [req-2602eece-b3e3-44ef-87cb-416b94d95dc7 req-871cb215-bdbd-45d9-9750-e3bd2d31ffca service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received event network-changed-14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.012184] env[62368]: DEBUG nova.compute.manager [req-2602eece-b3e3-44ef-87cb-416b94d95dc7 req-871cb215-bdbd-45d9-9750-e3bd2d31ffca service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing instance network info cache due to event network-changed-14840bf2-7da7-4d91-aa96-02a1771fb7d1. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 959.012457] env[62368]: DEBUG oslo_concurrency.lockutils [req-2602eece-b3e3-44ef-87cb-416b94d95dc7 req-871cb215-bdbd-45d9-9750-e3bd2d31ffca service nova] Acquiring lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.012617] env[62368]: DEBUG oslo_concurrency.lockutils [req-2602eece-b3e3-44ef-87cb-416b94d95dc7 req-871cb215-bdbd-45d9-9750-e3bd2d31ffca service nova] Acquired lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.012786] env[62368]: DEBUG nova.network.neutron [req-2602eece-b3e3-44ef-87cb-416b94d95dc7 req-871cb215-bdbd-45d9-9750-e3bd2d31ffca service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing network info cache for port 14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.348032] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.348380] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.432509] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.432509] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e99a37-54c4-3caa-6d86-f7405ce52677" [ 959.432509] env[62368]: _type = "HttpNfcLease" [ 959.432509] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.439655] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198875, 'name': Rename_Task, 'duration_secs': 0.316119} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.440410] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 959.440635] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-120be545-06d5-4c2e-9ca8-e84fe5412a0e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.449275] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 959.449275] env[62368]: value = "task-1198876" [ 959.449275] env[62368]: _type = "Task" [ 959.449275] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.458433] env[62368]: DEBUG nova.network.neutron [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance_info_cache with network_info: [{"id": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "address": "fa:16:3e:b0:23:0b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd229b53b-0e", "ovs_interfaceid": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.466650] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.490860] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5d2c7f-8064-45d3-99fe-cc828b98588b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.500576] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44e7513-21c4-4443-b640-fe81bd0c3e99 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.544575] env[62368]: DEBUG nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.545286] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.545698] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.545946] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.546171] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.546364] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.546553] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.546851] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.547032] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.547286] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.547533] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.547772] env[62368]: DEBUG nova.virt.hardware [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.549353] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6e3243-6d1f-4d47-bfa8-ec1bafa48b04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.552984] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56cbea3-41f8-4943-9a03-3448c272a312 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.565122] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f54903a-5631-4def-9def-0210feb0c1a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.570271] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afcf03b-335a-4eb9-9b1d-83858145e439 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.585394] env[62368]: DEBUG nova.compute.provider_tree [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.617365] env[62368]: DEBUG nova.network.neutron [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Successfully updated port: d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 959.811012] env[62368]: DEBUG nova.network.neutron [req-2602eece-b3e3-44ef-87cb-416b94d95dc7 req-871cb215-bdbd-45d9-9750-e3bd2d31ffca service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updated VIF entry in instance network info cache for port 14840bf2-7da7-4d91-aa96-02a1771fb7d1. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 959.811864] env[62368]: DEBUG nova.network.neutron [req-2602eece-b3e3-44ef-87cb-416b94d95dc7 req-871cb215-bdbd-45d9-9750-e3bd2d31ffca service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.824483] env[62368]: DEBUG nova.compute.manager [req-4e256721-3077-4280-ac11-384922fa65cb req-100993ba-73de-494f-b81f-f08ad2e23750 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Received event network-changed-d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.825237] env[62368]: DEBUG nova.compute.manager [req-4e256721-3077-4280-ac11-384922fa65cb req-100993ba-73de-494f-b81f-f08ad2e23750 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Refreshing instance network info cache due to event network-changed-d47a86a3-3fd6-48d2-8a80-e6b7f201bd76. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 959.825237] env[62368]: DEBUG oslo_concurrency.lockutils [req-4e256721-3077-4280-ac11-384922fa65cb req-100993ba-73de-494f-b81f-f08ad2e23750 service nova] Acquiring lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.825237] env[62368]: DEBUG oslo_concurrency.lockutils [req-4e256721-3077-4280-ac11-384922fa65cb req-100993ba-73de-494f-b81f-f08ad2e23750 service nova] Acquired lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.825963] env[62368]: DEBUG nova.network.neutron [req-4e256721-3077-4280-ac11-384922fa65cb req-100993ba-73de-494f-b81f-f08ad2e23750 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Refreshing network info cache for port d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.860341] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.863047] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 959.921307] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.921307] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e99a37-54c4-3caa-6d86-f7405ce52677" [ 959.921307] env[62368]: _type = "HttpNfcLease" [ 959.921307] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.962654] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.964179] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198876, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.021293] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.021794] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.099592] env[62368]: DEBUG nova.scheduler.client.report [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.120622] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Acquiring lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.121220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Acquired lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.121220] env[62368]: DEBUG nova.network.neutron [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 960.314805] env[62368]: DEBUG oslo_concurrency.lockutils [req-2602eece-b3e3-44ef-87cb-416b94d95dc7 req-871cb215-bdbd-45d9-9750-e3bd2d31ffca service nova] Releasing lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.378778] env[62368]: DEBUG nova.network.neutron [-] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.417849] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.417939] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquired lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.418082] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Forcefully refreshing network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 960.426657] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.426657] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e99a37-54c4-3caa-6d86-f7405ce52677" [ 960.426657] env[62368]: _type = "HttpNfcLease" [ 960.426657] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 960.426657] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 960.426657] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e99a37-54c4-3caa-6d86-f7405ce52677" [ 960.426657] env[62368]: _type = "HttpNfcLease" [ 960.426657] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 960.427040] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ac91cb-b911-4f3f-9922-b1237e4348a4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.437015] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524de3cf-90fd-0469-0e81-a3d77a843ea8/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 960.437015] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524de3cf-90fd-0469-0e81-a3d77a843ea8/disk-0.vmdk. {{(pid=62368) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 960.514725] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-777969bf-dd35-4611-b28f-f52670fb4a89 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.520463] env[62368]: DEBUG oslo_vmware.api [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198876, 'name': PowerOnVM_Task, 'duration_secs': 0.623853} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.521611] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 960.521815] env[62368]: INFO nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Took 9.26 seconds to spawn the instance on the hypervisor. [ 960.521989] env[62368]: DEBUG nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.522848] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c81c26b-8e8b-442a-b881-93d1201818f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.529956] env[62368]: DEBUG nova.compute.utils [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.604497] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.604940] env[62368]: DEBUG nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.654042] env[62368]: DEBUG nova.network.neutron [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 960.785440] env[62368]: DEBUG nova.network.neutron [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Updating instance_info_cache with network_info: [{"id": "d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c", "address": "fa:16:3e:2e:88:d0", "network": {"id": "4e121ffc-3515-4d77-8958-0cd1e568ebfa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1614832061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46f27452524540b596fa32b9999dc208", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4014180-3c", "ovs_interfaceid": "d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.888804] env[62368]: INFO nova.compute.manager [-] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Took 2.18 seconds to deallocate network for instance. [ 960.981476] env[62368]: DEBUG nova.network.neutron [req-4e256721-3077-4280-ac11-384922fa65cb req-100993ba-73de-494f-b81f-f08ad2e23750 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updated VIF entry in instance network info cache for port d47a86a3-3fd6-48d2-8a80-e6b7f201bd76. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.981634] env[62368]: DEBUG nova.network.neutron [req-4e256721-3077-4280-ac11-384922fa65cb req-100993ba-73de-494f-b81f-f08ad2e23750 service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updating instance_info_cache with network_info: [{"id": "d47a86a3-3fd6-48d2-8a80-e6b7f201bd76", "address": "fa:16:3e:21:07:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd47a86a3-3f", "ovs_interfaceid": "d47a86a3-3fd6-48d2-8a80-e6b7f201bd76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.989131] env[62368]: DEBUG nova.network.neutron [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Successfully updated port: d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.035079] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.042034] env[62368]: INFO nova.compute.manager [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Took 19.70 seconds to build instance. [ 961.110660] env[62368]: DEBUG nova.compute.utils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.112133] env[62368]: DEBUG nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 961.112358] env[62368]: DEBUG nova.network.neutron [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 961.155485] env[62368]: DEBUG nova.policy [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7dceb8c8f8b4c4dbdde4d097009d429', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b70fcc6664f47d7b55447210851c4cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.290983] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Releasing lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.291288] env[62368]: DEBUG nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Instance network_info: |[{"id": "d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c", "address": "fa:16:3e:2e:88:d0", "network": {"id": "4e121ffc-3515-4d77-8958-0cd1e568ebfa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1614832061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46f27452524540b596fa32b9999dc208", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4014180-3c", "ovs_interfaceid": "d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.291862] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:88:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2624812a-9f9c-461d-8b5f-79bea90c7ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.300383] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Creating folder: Project (46f27452524540b596fa32b9999dc208). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 961.303781] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f74fe6db-3fd3-4b69-a0e2-19a18ddb0724 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.321216] env[62368]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 961.321468] env[62368]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62368) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 961.322104] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Folder already exists: Project (46f27452524540b596fa32b9999dc208). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 961.322382] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Creating folder: Instances. Parent ref: group-v259859. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 961.322755] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b8136f8-9aaf-42cb-b6d8-3f8ca1e2fd15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.338718] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Created folder: Instances in parent group-v259859. [ 961.339073] env[62368]: DEBUG oslo.service.loopingcall [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.339288] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 961.339567] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cda58a37-f374-48a7-bd3b-d54b3a91ee09 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.366817] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.366817] env[62368]: value = "task-1198879" [ 961.366817] env[62368]: _type = "Task" [ 961.366817] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.380136] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198879, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.396470] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.396856] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.397374] env[62368]: DEBUG nova.objects.instance [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'resources' on Instance uuid 58157ab1-80a4-427c-812b-f6fde1f8db68 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.447671] env[62368]: DEBUG nova.network.neutron [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Successfully created port: 114fbe1a-b6e9-405a-9d63-241a22cb6560 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.487530] env[62368]: DEBUG oslo_concurrency.lockutils [req-4e256721-3077-4280-ac11-384922fa65cb req-100993ba-73de-494f-b81f-f08ad2e23750 service nova] Releasing lock "refresh_cache-1d5a93f0-1068-4514-b01c-011e95f48498" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.491836] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.495661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.495661] env[62368]: DEBUG nova.network.neutron [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.529333] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ace0df3-08ab-4a08-9c9f-f1908c11d5fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.560208] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c7399030-fc54-4a5a-a044-5640116c738e tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.230s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.561100] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance '72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2' progress to 0 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 961.615014] env[62368]: DEBUG nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.869199] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Updating instance_info_cache with network_info: [{"id": "bf67cd0c-ff81-451a-8e0e-6ee50bb77f78", "address": "fa:16:3e:90:7b:bf", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf67cd0c-ff", "ovs_interfaceid": "bf67cd0c-ff81-451a-8e0e-6ee50bb77f78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.886098] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198879, 'name': CreateVM_Task, 'duration_secs': 0.482937} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.887939] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 961.888886] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'guest_format': None, 'attachment_id': '5af1f9e6-f9c8-454c-8839-c012c8528734', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259862', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'name': 'volume-0923c8dd-6948-46a8-a7ce-88b165401227', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1b238e7-3636-47db-b86e-3fcfe917f705', 'attached_at': '', 'detached_at': '', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'serial': '0923c8dd-6948-46a8-a7ce-88b165401227'}, 'mount_device': '/dev/sda', 'boot_index': 0, 'disk_bus': None, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=62368) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 961.889182] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Root volume attach. Driver type: vmdk {{(pid=62368) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 961.890217] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30214537-0511-4c8b-98ec-47e8a4c54e4e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.900134] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498dde1d-114c-4380-8130-7c32f0262fc7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.912627] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa208fb-0144-46e5-9215-14e139010457 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.923943] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-f596d33a-c1df-4fd0-8cca-e3bcb411235e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.941757] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 961.941757] env[62368]: value = "task-1198880" [ 961.941757] env[62368]: _type = "Task" [ 961.941757] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.954920] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.059864] env[62368]: WARNING nova.network.neutron [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] 3ee164de-30af-4404-935a-a7f988d824b1 already exists in list: networks containing: ['3ee164de-30af-4404-935a-a7f988d824b1']. ignoring it [ 962.076164] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 962.076519] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd3c6723-41ba-4802-a696-fec072b10433 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.085479] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 962.085479] env[62368]: value = "task-1198881" [ 962.085479] env[62368]: _type = "Task" [ 962.085479] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.096128] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198881, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.111020] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.111299] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.111529] env[62368]: INFO nova.compute.manager [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Attaching volume cf7750c2-86cd-43ff-89ac-880138be3dc5 to /dev/sdb [ 962.157411] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f382418c-178b-4c2a-9cb9-9cbbee0f79ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.170560] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf04d157-22bc-493f-bf60-75049ee271e6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.195070] env[62368]: DEBUG nova.virt.block_device [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating existing volume attachment record: 2175cda4-5fb0-47d8-becc-a70f5b878253 {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 962.257748] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Completed reading data from the image iterator. {{(pid=62368) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 962.258111] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524de3cf-90fd-0469-0e81-a3d77a843ea8/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 962.259348] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15e3779-1c8c-40b5-982a-4b821030b37f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.272503] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524de3cf-90fd-0469-0e81-a3d77a843ea8/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 962.272503] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524de3cf-90fd-0469-0e81-a3d77a843ea8/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 962.272503] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-d66060d9-0421-44ad-88f0-7a07e672fadd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.312621] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47be2ea-9c9d-43d9-9a5d-7803c84e5ad7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.324916] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529be63c-9d06-4de8-a08b-905baa0ee7bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.378483] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d427dd08-b7e4-4722-8039-ea9a4a6e06f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.381933] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Releasing lock "refresh_cache-b79f0e79-9e3a-47c7-9949-8743601ec6c5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.382196] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Updated the network info_cache for instance {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 962.382496] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.383476] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.383760] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.384040] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.384349] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.384509] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.384699] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 962.384958] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.390694] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc11f1c-c28e-4134-904f-9016325fc19e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.408285] env[62368]: DEBUG nova.compute.provider_tree [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.458568] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 40%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.601137] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198881, 'name': PowerOffVM_Task, 'duration_secs': 0.495815} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.601534] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 962.601746] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance '72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2' progress to 17 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 962.612278] env[62368]: DEBUG oslo_vmware.rw_handles [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524de3cf-90fd-0469-0e81-a3d77a843ea8/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 962.612744] env[62368]: INFO nova.virt.vmwareapi.images [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Downloaded image file data bc3d2926-26a5-47b1-b7eb-6a838c226d70 [ 962.613540] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83de412-e5f9-40be-888a-d4c99dc66b13 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.636249] env[62368]: DEBUG nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.638066] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-deb56c85-8b71-4c5d-8850-5429bade400f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.655697] env[62368]: DEBUG nova.network.neutron [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1b97d93-2fbd-4ff2-948f-6f197d6ed77f", "address": "fa:16:3e:26:cf:d1", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1b97d93-2f", "ovs_interfaceid": "d1b97d93-2fbd-4ff2-948f-6f197d6ed77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.671757] env[62368]: INFO nova.virt.vmwareapi.images [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] The imported VM was unregistered [ 962.674841] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Caching image {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 962.675983] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Creating directory with path [datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.678445] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.678816] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.679089] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.679483] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.679483] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.679652] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.679964] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.680209] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.680431] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.680690] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.681115] env[62368]: DEBUG nova.virt.hardware [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.682174] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f540357-dcbf-4fa8-86c3-da655994b73a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.686031] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfc8762-fc25-49ec-825e-0b0cb91baea9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.699056] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4b0d8d-e0a7-407f-87a0-e1167894203e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.718922] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Created directory with path [datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.719175] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf/OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf.vmdk to [datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70/bc3d2926-26a5-47b1-b7eb-6a838c226d70.vmdk. {{(pid=62368) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 962.719458] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-5a7424ef-ea95-4517-841e-fd1a40c3d54a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.731519] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 962.731519] env[62368]: value = "task-1198886" [ 962.731519] env[62368]: _type = "Task" [ 962.731519] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.732772] env[62368]: DEBUG nova.compute.manager [req-3f0f1331-5820-49d5-a1f4-b6259ff5ee36 req-ebc2a487-d557-4929-b261-4a6e8e39f74e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received event network-changed-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.732967] env[62368]: DEBUG nova.compute.manager [req-3f0f1331-5820-49d5-a1f4-b6259ff5ee36 req-ebc2a487-d557-4929-b261-4a6e8e39f74e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Refreshing instance network info cache due to event network-changed-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.733496] env[62368]: DEBUG oslo_concurrency.lockutils [req-3f0f1331-5820-49d5-a1f4-b6259ff5ee36 req-ebc2a487-d557-4929-b261-4a6e8e39f74e service nova] Acquiring lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.733670] env[62368]: DEBUG oslo_concurrency.lockutils [req-3f0f1331-5820-49d5-a1f4-b6259ff5ee36 req-ebc2a487-d557-4929-b261-4a6e8e39f74e service nova] Acquired lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.733854] env[62368]: DEBUG nova.network.neutron [req-3f0f1331-5820-49d5-a1f4-b6259ff5ee36 req-ebc2a487-d557-4929-b261-4a6e8e39f74e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Refreshing network info cache for port 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.751716] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198886, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.763512] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Received event network-vif-plugged-d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.763794] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Acquiring lock "c1b238e7-3636-47db-b86e-3fcfe917f705-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.764186] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Lock "c1b238e7-3636-47db-b86e-3fcfe917f705-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.764426] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Lock "c1b238e7-3636-47db-b86e-3fcfe917f705-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.764610] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] No waiting events found dispatching network-vif-plugged-d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 962.764787] env[62368]: WARNING nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Received unexpected event network-vif-plugged-d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c for instance with vm_state building and task_state spawning. [ 962.764958] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Received event network-changed-d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.765141] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Refreshing instance network info cache due to event network-changed-d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.765324] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Acquiring lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.765463] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Acquired lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.765656] env[62368]: DEBUG nova.network.neutron [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Refreshing network info cache for port d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.889317] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.913944] env[62368]: DEBUG nova.scheduler.client.report [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.957833] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 54%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.109764] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.110036] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.110199] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.110427] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.110799] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.110973] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.111212] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.111384] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.111563] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.111828] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.112128] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.117932] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e2c5454-0692-42fd-a733-d50c5b2e81a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.139811] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 963.139811] env[62368]: value = "task-1198887" [ 963.139811] env[62368]: _type = "Task" [ 963.139811] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.151119] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198887, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.161355] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.162169] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.162368] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.163756] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5679455f-e94f-48c2-ab78-7630b764931e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.187625] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.188293] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.188455] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.188687] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.188837] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.189035] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.189262] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.189439] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.189633] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.189846] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.189998] env[62368]: DEBUG nova.virt.hardware [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.197378] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Reconfiguring VM to attach interface {{(pid=62368) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 963.198010] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22baff5d-078a-405d-9433-884ada4ccd59 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.223348] env[62368]: DEBUG oslo_vmware.api [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 963.223348] env[62368]: value = "task-1198888" [ 963.223348] env[62368]: _type = "Task" [ 963.223348] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.236955] env[62368]: DEBUG oslo_vmware.api [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198888, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.256605] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198886, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.423092] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.026s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.426596] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.537s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.427051] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.427248] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 963.428533] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487acc1b-fef8-4e7a-9ac3-b1665cd30ed9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.454435] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bd065a-e87b-43d6-9274-ffcb5f7d7042 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.461904] env[62368]: INFO nova.scheduler.client.report [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Deleted allocations for instance 58157ab1-80a4-427c-812b-f6fde1f8db68 [ 963.475928] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 67%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.487613] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad6d94c-2dc7-4b34-bd13-21b5a2ab2ef0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.499293] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d7b67f-1ccb-473e-9aae-d8402788030e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.543077] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179806MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 963.543308] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.543428] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.549656] env[62368]: DEBUG nova.network.neutron [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Updated VIF entry in instance network info cache for port d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 963.550132] env[62368]: DEBUG nova.network.neutron [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Updating instance_info_cache with network_info: [{"id": "d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c", "address": "fa:16:3e:2e:88:d0", "network": {"id": "4e121ffc-3515-4d77-8958-0cd1e568ebfa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1614832061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46f27452524540b596fa32b9999dc208", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4014180-3c", "ovs_interfaceid": "d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.622063] env[62368]: DEBUG nova.network.neutron [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Successfully updated port: 114fbe1a-b6e9-405a-9d63-241a22cb6560 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.659913] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198887, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.737793] env[62368]: DEBUG oslo_vmware.api [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198888, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.750350] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198886, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.789641] env[62368]: DEBUG nova.network.neutron [req-3f0f1331-5820-49d5-a1f4-b6259ff5ee36 req-ebc2a487-d557-4929-b261-4a6e8e39f74e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updated VIF entry in instance network info cache for port 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 963.790215] env[62368]: DEBUG nova.network.neutron [req-3f0f1331-5820-49d5-a1f4-b6259ff5ee36 req-ebc2a487-d557-4929-b261-4a6e8e39f74e service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [{"id": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "address": "fa:16:3e:a8:8e:88", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15eed45d-e0", "ovs_interfaceid": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.955415] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 81%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.974413] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85e2363a-4616-4861-a018-8b73f8810b01 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "58157ab1-80a4-427c-812b-f6fde1f8db68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.479s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.053320] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Releasing lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.054057] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Received event network-vif-deleted-46fba643-185a-4fb2-ab24-dd4350aa5ac4 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.054650] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received event network-vif-plugged-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.054650] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Acquiring lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.054820] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.055017] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.055199] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] No waiting events found dispatching network-vif-plugged-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 964.055379] env[62368]: WARNING nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received unexpected event network-vif-plugged-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f for instance with vm_state active and task_state None. [ 964.055634] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received event network-changed-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.055846] env[62368]: DEBUG nova.compute.manager [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing instance network info cache due to event network-changed-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.056151] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Acquiring lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.056357] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Acquired lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.056547] env[62368]: DEBUG nova.network.neutron [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Refreshing network info cache for port d1b97d93-2fbd-4ff2-948f-6f197d6ed77f {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 964.127118] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-84bb7208-670f-4e1d-862a-a480f7db0f42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.127290] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-84bb7208-670f-4e1d-862a-a480f7db0f42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.127436] env[62368]: DEBUG nova.network.neutron [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.148765] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.149273] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.149648] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.150136] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.150457] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.157682] env[62368]: INFO nova.compute.manager [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Terminating instance [ 964.164392] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198887, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.165391] env[62368]: DEBUG nova.compute.manager [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.165719] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 964.166619] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaa29cb-7f67-4fa3-b4bb-c3fadcdfec3d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.178509] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 964.178858] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6601b385-48d3-4e4e-a221-54557580d6a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.191424] env[62368]: DEBUG oslo_vmware.api [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 964.191424] env[62368]: value = "task-1198889" [ 964.191424] env[62368]: _type = "Task" [ 964.191424] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.204514] env[62368]: DEBUG oslo_vmware.api [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.237936] env[62368]: DEBUG oslo_vmware.api [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198888, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.254255] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198886, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.294538] env[62368]: DEBUG oslo_concurrency.lockutils [req-3f0f1331-5820-49d5-a1f4-b6259ff5ee36 req-ebc2a487-d557-4929-b261-4a6e8e39f74e service nova] Releasing lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.456249] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 95%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.573250] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Applying migration context for instance 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2 as it has an incoming, in-progress migration e3cd6a40-1f21-4511-adbd-66cdf0420c64. Migration status is migrating {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 964.575672] env[62368]: INFO nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating resource usage from migration e3cd6a40-1f21-4511-adbd-66cdf0420c64 [ 964.604041] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b79f0e79-9e3a-47c7-9949-8743601ec6c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.604162] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 2a46c954-449e-4d62-be80-add1040ed4c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.604289] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b87094da-6258-469e-ab37-5557955ad3a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.604413] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 69ddb565-6c79-44e5-a7d1-d339ab426fae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.604555] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a5cbadbd-20dd-4514-8867-20243af5db0c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.604694] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 3e09be20-ce79-4bfa-bf84-2877f4e534e6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.604810] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 5633292c-9a74-4c2d-893b-6823f4a587a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.604924] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance e427d283-1f8f-4c7a-869c-4cd48a595470 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.605044] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance e6dd9825-f841-49ac-a75a-83d9de3858ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.605160] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance eea21546-fbbf-4440-829c-8583c4ccabb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.605274] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance fa637a2e-a113-4d23-8924-728ba75f202c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.605382] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 1d5a93f0-1068-4514-b01c-011e95f48498 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.605494] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance cf3601d5-2956-4301-a7cc-bc2d33bf3790 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.605673] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance c1b238e7-3636-47db-b86e-3fcfe917f705 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.605806] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Migration e3cd6a40-1f21-4511-adbd-66cdf0420c64 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 964.605920] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.606045] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 84bb7208-670f-4e1d-862a-a480f7db0f42 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.606272] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 17 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 964.606415] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3840MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=17 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 964.653827] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198887, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.673190] env[62368]: DEBUG nova.network.neutron [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 964.707297] env[62368]: DEBUG oslo_vmware.api [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198889, 'name': PowerOffVM_Task, 'duration_secs': 0.394295} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.707566] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 964.707892] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 964.707994] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da61c078-a6db-40be-bfc2-18c3ddb16bfe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.745874] env[62368]: DEBUG oslo_vmware.api [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198888, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.767479] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198886, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.804265] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 964.804546] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 964.804748] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleting the datastore file [datastore2] cf3601d5-2956-4301-a7cc-bc2d33bf3790 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.805111] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8992d18b-e697-4395-b80e-626adb701c04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.811432] env[62368]: DEBUG nova.compute.manager [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Received event network-vif-plugged-114fbe1a-b6e9-405a-9d63-241a22cb6560 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.811432] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] Acquiring lock "84bb7208-670f-4e1d-862a-a480f7db0f42-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.811432] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.811432] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.811432] env[62368]: DEBUG nova.compute.manager [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] No waiting events found dispatching network-vif-plugged-114fbe1a-b6e9-405a-9d63-241a22cb6560 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 964.811432] env[62368]: WARNING nova.compute.manager [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Received unexpected event network-vif-plugged-114fbe1a-b6e9-405a-9d63-241a22cb6560 for instance with vm_state building and task_state spawning. [ 964.811432] env[62368]: DEBUG nova.compute.manager [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Received event network-changed-114fbe1a-b6e9-405a-9d63-241a22cb6560 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.811731] env[62368]: DEBUG nova.compute.manager [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Refreshing instance network info cache due to event network-changed-114fbe1a-b6e9-405a-9d63-241a22cb6560. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.811731] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] Acquiring lock "refresh_cache-84bb7208-670f-4e1d-862a-a480f7db0f42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.821838] env[62368]: DEBUG oslo_vmware.api [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 964.821838] env[62368]: value = "task-1198891" [ 964.821838] env[62368]: _type = "Task" [ 964.821838] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.834967] env[62368]: DEBUG oslo_vmware.api [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198891, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.897464] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830da08e-f1b1-40bb-9a74-42adb4c1f7c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.906079] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed7b589-c16f-4d1f-8941-ce044b39e492 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.944568] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f7dcff-71f0-4d25-9f81-47903b4be26d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.959995] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c3f62f-d0be-43a9-9443-e674837b4652 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.964242] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 97%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.979065] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 964.984663] env[62368]: DEBUG nova.network.neutron [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updated VIF entry in instance network info cache for port d1b97d93-2fbd-4ff2-948f-6f197d6ed77f. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 964.985227] env[62368]: DEBUG nova.network.neutron [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d1b97d93-2fbd-4ff2-948f-6f197d6ed77f", "address": "fa:16:3e:26:cf:d1", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1b97d93-2f", "ovs_interfaceid": "d1b97d93-2fbd-4ff2-948f-6f197d6ed77f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.990550] env[62368]: DEBUG nova.network.neutron [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Updating instance_info_cache with network_info: [{"id": "114fbe1a-b6e9-405a-9d63-241a22cb6560", "address": "fa:16:3e:1a:42:37", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114fbe1a-b6", "ovs_interfaceid": "114fbe1a-b6e9-405a-9d63-241a22cb6560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.158295] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198887, 'name': ReconfigVM_Task, 'duration_secs': 1.94519} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.158860] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance '72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2' progress to 33 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 965.240201] env[62368]: DEBUG oslo_vmware.api [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198888, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.258246] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198886, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.335509] env[62368]: DEBUG oslo_vmware.api [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198891, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.417799} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.335911] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.336178] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 965.336434] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 965.336672] env[62368]: INFO nova.compute.manager [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Took 1.17 seconds to destroy the instance on the hypervisor. [ 965.337023] env[62368]: DEBUG oslo.service.loopingcall [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.337291] env[62368]: DEBUG nova.compute.manager [-] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.337373] env[62368]: DEBUG nova.network.neutron [-] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 965.459023] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 97%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.493370] env[62368]: DEBUG oslo_concurrency.lockutils [req-bf90e16e-8c33-4b4d-b7cb-3de8518a3e34 req-f9313df4-7ad0-40d9-bd72-c0fa26623c44 service nova] Releasing lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.496618] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-84bb7208-670f-4e1d-862a-a480f7db0f42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.496972] env[62368]: DEBUG nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Instance network_info: |[{"id": "114fbe1a-b6e9-405a-9d63-241a22cb6560", "address": "fa:16:3e:1a:42:37", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114fbe1a-b6", "ovs_interfaceid": "114fbe1a-b6e9-405a-9d63-241a22cb6560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.497503] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] Acquired lock "refresh_cache-84bb7208-670f-4e1d-862a-a480f7db0f42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.497695] env[62368]: DEBUG nova.network.neutron [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Refreshing network info cache for port 114fbe1a-b6e9-405a-9d63-241a22cb6560 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 965.498955] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:42:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '114fbe1a-b6e9-405a-9d63-241a22cb6560', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.507189] env[62368]: DEBUG oslo.service.loopingcall [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.507865] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 965.507975] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ef2e4b0-2b5a-475a-8abd-02290fdb5dc5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.526692] env[62368]: ERROR nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [req-b421c5a0-5a4b-4176-b5e6-821905ffb674] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b421c5a0-5a4b-4176-b5e6-821905ffb674"}]} [ 965.531777] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.531777] env[62368]: value = "task-1198892" [ 965.531777] env[62368]: _type = "Task" [ 965.531777] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.541703] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198892, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.560395] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 965.576572] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 965.576735] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 965.595101] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 965.620944] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 965.667584] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.667584] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.667584] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.667764] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.668493] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.668493] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.668493] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.668493] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.668693] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.668725] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.669140] env[62368]: DEBUG nova.virt.hardware [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.677310] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 965.680923] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d33d404-c6cc-430c-948e-3d6cf3c6994c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.706726] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 965.706726] env[62368]: value = "task-1198893" [ 965.706726] env[62368]: _type = "Task" [ 965.706726] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.721997] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198893, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.743760] env[62368]: DEBUG oslo_vmware.api [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198888, 'name': ReconfigVM_Task, 'duration_secs': 2.393004} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.744304] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.744520] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Reconfigured VM to attach interface {{(pid=62368) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 965.761466] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198886, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.686658} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.765038] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf/OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf.vmdk to [datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70/bc3d2926-26a5-47b1-b7eb-6a838c226d70.vmdk. [ 965.765296] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Cleaning up location [datastore1] OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 965.765541] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_7d59e48b-b2dd-4610-8cd0-91f34f2d7ecf {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.766068] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3aad6ff1-7034-4dc5-b853-33c7bda69875 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.774852] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 965.774852] env[62368]: value = "task-1198894" [ 965.774852] env[62368]: _type = "Task" [ 965.774852] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.786592] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198894, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.911589] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e419874f-f131-4f90-a89f-7f39d3fb29d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.921032] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85ec9c5-6a9b-407a-9c35-47af04f1455a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.955848] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69537124-012a-43f1-90de-123da4eac612 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.967376] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task} progress is 98%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.969025] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197525c8-f2f3-4dea-b9e8-cf72eaa1ce5e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.986794] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.044133] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198892, 'name': CreateVM_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.216360] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198893, 'name': ReconfigVM_Task, 'duration_secs': 0.497197} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.216633] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 966.216994] env[62368]: DEBUG nova.network.neutron [-] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.221455] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09849d89-ed71-4ae4-88fb-6efb1b1167a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.247824] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2/72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.251864] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2a93941-484f-4c70-ab23-04745f0c0ac5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.267538] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5f645390-bfb5-4e33-8417-53a6f5be9766 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-5633292c-9a74-4c2d-893b-6823f4a587a2-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.542s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.276488] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 966.276488] env[62368]: value = "task-1198895" [ 966.276488] env[62368]: _type = "Task" [ 966.276488] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.289798] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198894, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114092} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.292906] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.293091] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70/bc3d2926-26a5-47b1-b7eb-6a838c226d70.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.293345] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70/bc3d2926-26a5-47b1-b7eb-6a838c226d70.vmdk to [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6/eea21546-fbbf-4440-829c-8583c4ccabb6.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 966.293892] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198895, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.294116] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5f299bd-bd7a-4ae9-8425-688054b870f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.302033] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 966.302033] env[62368]: value = "task-1198896" [ 966.302033] env[62368]: _type = "Task" [ 966.302033] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.311199] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.332462] env[62368]: DEBUG nova.network.neutron [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Updated VIF entry in instance network info cache for port 114fbe1a-b6e9-405a-9d63-241a22cb6560. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 966.332932] env[62368]: DEBUG nova.network.neutron [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Updating instance_info_cache with network_info: [{"id": "114fbe1a-b6e9-405a-9d63-241a22cb6560", "address": "fa:16:3e:1a:42:37", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114fbe1a-b6", "ovs_interfaceid": "114fbe1a-b6e9-405a-9d63-241a22cb6560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.462580] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198880, 'name': RelocateVM_Task, 'duration_secs': 4.39096} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.463714] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 966.463714] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259862', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'name': 'volume-0923c8dd-6948-46a8-a7ce-88b165401227', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1b238e7-3636-47db-b86e-3fcfe917f705', 'attached_at': '', 'detached_at': '', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'serial': '0923c8dd-6948-46a8-a7ce-88b165401227'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 966.464153] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec2fda0-f1b1-4498-85e2-829978dc76a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.482217] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d772a733-4b54-4e7d-a9c8-0b848956abc6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.488899] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.510452] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] volume-0923c8dd-6948-46a8-a7ce-88b165401227/volume-0923c8dd-6948-46a8-a7ce-88b165401227.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.511956] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 966.512157] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.969s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.512380] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b48455c0-d7e4-42be-80bb-9fc5df953f9d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.532931] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 966.532931] env[62368]: value = "task-1198897" [ 966.532931] env[62368]: _type = "Task" [ 966.532931] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.548769] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198897, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.549036] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198892, 'name': CreateVM_Task, 'duration_secs': 0.813883} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.549216] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 966.550025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.550108] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.550494] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.550773] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fbc3b03-0bd8-430f-950c-6970e1e7f4aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.557087] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 966.557087] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5238740e-5678-b645-1fd4-a77c39add2a2" [ 966.557087] env[62368]: _type = "Task" [ 966.557087] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.566277] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5238740e-5678-b645-1fd4-a77c39add2a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.722652] env[62368]: INFO nova.compute.manager [-] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Took 1.39 seconds to deallocate network for instance. [ 966.792702] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198895, 'name': ReconfigVM_Task, 'duration_secs': 0.376499} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.793076] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2/72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.793448] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance '72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2' progress to 50 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 966.815644] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198896, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.836857] env[62368]: DEBUG oslo_concurrency.lockutils [req-ec8fa027-eac1-4032-ae81-e26d623861e3 req-9db5cd79-cf01-41ce-b785-dca9f6834496 service nova] Releasing lock "refresh_cache-84bb7208-670f-4e1d-862a-a480f7db0f42" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.954305] env[62368]: DEBUG nova.compute.manager [req-21202f0f-1fcf-4992-83b6-5410de6f8472 req-1ee3b541-0688-4457-9ab0-59e8ae7c0839 service nova] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Received event network-vif-deleted-f1a74431-d8b5-4107-9e82-02521ef5e35b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.046120] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198897, 'name': ReconfigVM_Task, 'duration_secs': 0.328696} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.046649] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Reconfigured VM instance instance-00000061 to attach disk [datastore1] volume-0923c8dd-6948-46a8-a7ce-88b165401227/volume-0923c8dd-6948-46a8-a7ce-88b165401227.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.051867] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3cb7c96-c1ba-4bdb-ab4a-db46b60bf506 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.074900] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5238740e-5678-b645-1fd4-a77c39add2a2, 'name': SearchDatastore_Task, 'duration_secs': 0.011293} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.076615] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.076992] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.077248] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.077713] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.077713] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.077997] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 967.077997] env[62368]: value = "task-1198899" [ 967.077997] env[62368]: _type = "Task" [ 967.077997] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.078234] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c495a803-2dba-47b3-b496-f76c536e42f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.090824] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198899, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.101335] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.101335] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 967.102111] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f050033d-7e0c-4a6b-a0e5-90d88208815d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.111074] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 967.111074] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520442c5-012f-5e77-c59a-40cbefbfde37" [ 967.111074] env[62368]: _type = "Task" [ 967.111074] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.123885] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520442c5-012f-5e77-c59a-40cbefbfde37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.229665] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.229807] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.230114] env[62368]: DEBUG nova.objects.instance [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lazy-loading 'resources' on Instance uuid cf3601d5-2956-4301-a7cc-bc2d33bf3790 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.255309] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.255485] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.300408] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4424f50c-cde8-451a-abac-05c896bfadfa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.313033] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198896, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.327389] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c97ef34-51e4-412c-8201-87f3627aad76 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.358819] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance '72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2' progress to 67 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 967.593510] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198899, 'name': ReconfigVM_Task, 'duration_secs': 0.14495} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.593854] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259862', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'name': 'volume-0923c8dd-6948-46a8-a7ce-88b165401227', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1b238e7-3636-47db-b86e-3fcfe917f705', 'attached_at': '', 'detached_at': '', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'serial': '0923c8dd-6948-46a8-a7ce-88b165401227'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 967.594592] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c8a60c6-d658-4298-befc-5723321d2a19 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.605582] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 967.605582] env[62368]: value = "task-1198900" [ 967.605582] env[62368]: _type = "Task" [ 967.605582] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.619019] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198900, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.625242] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520442c5-012f-5e77-c59a-40cbefbfde37, 'name': SearchDatastore_Task, 'duration_secs': 0.086597} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.626140] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d19687dc-d718-4ac8-b127-7a9f1cfb9b0b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.632224] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 967.632224] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52cf660e-0c68-9d3a-95ce-7498bd2a5c66" [ 967.632224] env[62368]: _type = "Task" [ 967.632224] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.641949] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cf660e-0c68-9d3a-95ce-7498bd2a5c66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.759416] env[62368]: DEBUG nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.814763] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198896, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.929779] env[62368]: DEBUG nova.network.neutron [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Port d229b53b-0e94-41c9-b8e9-0b0eaa366d3c binding to destination host cpu-1 is already ACTIVE {{(pid=62368) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 968.005796] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b61cc8-0214-4674-ab48-5c09ea3c4f91 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.014705] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463ffc64-8159-46e0-a21c-bb3be556be13 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.046517] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d022db3e-64c9-46ec-8c93-0a0fe2da92de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.055387] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc905119-3ea9-49e0-8acd-455ab5259895 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.070330] env[62368]: DEBUG nova.compute.provider_tree [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.116148] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198900, 'name': Rename_Task, 'duration_secs': 0.164287} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.116491] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 968.116740] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ee97f4d-2e67-4e30-97eb-a6ee4c9d3116 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.126601] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 968.126601] env[62368]: value = "task-1198901" [ 968.126601] env[62368]: _type = "Task" [ 968.126601] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.137162] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.146724] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cf660e-0c68-9d3a-95ce-7498bd2a5c66, 'name': SearchDatastore_Task, 'duration_secs': 0.087857} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.147035] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.147326] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 84bb7208-670f-4e1d-862a-a480f7db0f42/84bb7208-670f-4e1d-862a-a480f7db0f42.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 968.147610] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6aee9265-b358-476a-90c4-dff48a173be6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.168442] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 968.168442] env[62368]: value = "task-1198902" [ 968.168442] env[62368]: _type = "Task" [ 968.168442] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.178777] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "interface-5633292c-9a74-4c2d-893b-6823f4a587a2-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.179084] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-5633292c-9a74-4c2d-893b-6823f4a587a2-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.181455] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198902, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.281769] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.315706] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198896, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.573362] env[62368]: DEBUG nova.scheduler.client.report [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.636960] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198901, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.678770] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198902, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.682628] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.682888] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.683827] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3036f2b-264a-406d-a183-842a9a9adc4a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.704448] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e18ee9-50d6-4195-8fc2-713b18f92855 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.733861] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Reconfiguring VM to detach interface {{(pid=62368) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 968.734246] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d868a3e0-e8dd-4f6a-8f7d-bd80dc5eb47c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.754995] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 968.754995] env[62368]: value = "task-1198903" [ 968.754995] env[62368]: _type = "Task" [ 968.754995] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.767654] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.816241] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198896, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.958549] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.958964] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.958964] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.079297] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.081809] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.800s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.083480] env[62368]: INFO nova.compute.claims [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.104615] env[62368]: INFO nova.scheduler.client.report [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted allocations for instance cf3601d5-2956-4301-a7cc-bc2d33bf3790 [ 969.137770] env[62368]: DEBUG oslo_vmware.api [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1198901, 'name': PowerOnVM_Task, 'duration_secs': 0.818423} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.138852] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 969.139088] env[62368]: INFO nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Took 9.59 seconds to spawn the instance on the hypervisor. [ 969.139277] env[62368]: DEBUG nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.140140] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400ca4ea-cf90-49bc-9d5b-b576d8564233 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.180186] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198902, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.266032] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.315552] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198896, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.521354} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.315835] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bc3d2926-26a5-47b1-b7eb-6a838c226d70/bc3d2926-26a5-47b1-b7eb-6a838c226d70.vmdk to [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6/eea21546-fbbf-4440-829c-8583c4ccabb6.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 969.316636] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818ef6a4-33db-43e6-a1fd-d7c38df418f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.337937] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6/eea21546-fbbf-4440-829c-8583c4ccabb6.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.338534] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bea2f1b5-1a96-4621-8b5f-aee270f7d1ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.358689] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 969.358689] env[62368]: value = "task-1198904" [ 969.358689] env[62368]: _type = "Task" [ 969.358689] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.367259] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198904, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.612942] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7cf979f4-29ba-4138-aa7b-3636d848e1e1 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "cf3601d5-2956-4301-a7cc-bc2d33bf3790" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.464s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.660627] env[62368]: INFO nova.compute.manager [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Took 25.42 seconds to build instance. [ 969.681573] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198902, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.062851} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.681820] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 84bb7208-670f-4e1d-862a-a480f7db0f42/84bb7208-670f-4e1d-862a-a480f7db0f42.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 969.682058] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.682521] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f016242b-dc0d-4750-aede-cb04d49a0d14 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.689452] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 969.689452] env[62368]: value = "task-1198905" [ 969.689452] env[62368]: _type = "Task" [ 969.689452] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.697321] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198905, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.766390] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.869994] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198904, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.008109] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.008376] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.008597] env[62368]: DEBUG nova.network.neutron [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 970.164016] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c5794f3f-cd17-4a86-8f15-22be21dc6037 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "c1b238e7-3636-47db-b86e-3fcfe917f705" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.929s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.199402] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198905, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070195} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.201825] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.202825] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cc4a7e-f7a2-47f0-b950-1a66ce49253e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.227147] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 84bb7208-670f-4e1d-862a-a480f7db0f42/84bb7208-670f-4e1d-862a-a480f7db0f42.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.230100] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5e599e2-b931-4887-841d-17a58b467eb9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.252125] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 970.252125] env[62368]: value = "task-1198906" [ 970.252125] env[62368]: _type = "Task" [ 970.252125] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.264735] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198906, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.270318] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.367789] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e53da9-5116-4b14-ba8d-ecc6cbdfa7e3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.376077] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198904, 'name': ReconfigVM_Task, 'duration_secs': 0.999683} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.378300] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Reconfigured VM instance instance-00000040 to attach disk [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6/eea21546-fbbf-4440-829c-8583c4ccabb6.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.379083] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a07ef74-caa5-4a61-8a4b-f77b00331b9b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.381670] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43e483d-fc5c-4feb-ac78-b06a40d9cbb9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.413468] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "e427d283-1f8f-4c7a-869c-4cd48a595470" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.413892] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "e427d283-1f8f-4c7a-869c-4cd48a595470" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.414208] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "e427d283-1f8f-4c7a-869c-4cd48a595470-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.414511] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "e427d283-1f8f-4c7a-869c-4cd48a595470-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.414770] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "e427d283-1f8f-4c7a-869c-4cd48a595470-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.419066] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41c1840-c90f-486b-8f42-82cd2d880ebc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.421765] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 970.421765] env[62368]: value = "task-1198907" [ 970.421765] env[62368]: _type = "Task" [ 970.421765] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.422250] env[62368]: INFO nova.compute.manager [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Terminating instance [ 970.429864] env[62368]: DEBUG nova.compute.manager [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.430193] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 970.430925] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe77105-9ebe-47b0-8d44-7bee4996a6bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.434884] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e3724d-46a9-4e48-94ad-ae8ab506aed6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.442603] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198907, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.446987] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 970.454927] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d46cebbc-fb65-43db-bc44-818373a011bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.456749] env[62368]: DEBUG nova.compute.provider_tree [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.463728] env[62368]: DEBUG oslo_vmware.api [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 970.463728] env[62368]: value = "task-1198908" [ 970.463728] env[62368]: _type = "Task" [ 970.463728] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.473691] env[62368]: DEBUG oslo_vmware.api [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.757892] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 970.758217] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259876', 'volume_id': 'cf7750c2-86cd-43ff-89ac-880138be3dc5', 'name': 'volume-cf7750c2-86cd-43ff-89ac-880138be3dc5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e09be20-ce79-4bfa-bf84-2877f4e534e6', 'attached_at': '', 'detached_at': '', 'volume_id': 'cf7750c2-86cd-43ff-89ac-880138be3dc5', 'serial': 'cf7750c2-86cd-43ff-89ac-880138be3dc5'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 970.759080] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666ddc7f-c8d7-4f7f-866b-8da11ff08347 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.769060] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198906, 'name': ReconfigVM_Task, 'duration_secs': 0.281389} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.769307] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 84bb7208-670f-4e1d-862a-a480f7db0f42/84bb7208-670f-4e1d-862a-a480f7db0f42.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.769952] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-993d4764-314c-4a2a-a718-35441a6b3570 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.786524] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.787483] env[62368]: DEBUG nova.network.neutron [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance_info_cache with network_info: [{"id": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "address": "fa:16:3e:b0:23:0b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd229b53b-0e", "ovs_interfaceid": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.789462] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120c58fe-0554-41ea-9d65-7d5b5534df57 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.794777] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 970.794777] env[62368]: value = "task-1198909" [ 970.794777] env[62368]: _type = "Task" [ 970.794777] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.821579] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] volume-cf7750c2-86cd-43ff-89ac-880138be3dc5/volume-cf7750c2-86cd-43ff-89ac-880138be3dc5.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.828375] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26a33036-9d21-4f9a-8e24-79a2ee0ca9cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.843547] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198909, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.850485] env[62368]: DEBUG oslo_vmware.api [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 970.850485] env[62368]: value = "task-1198910" [ 970.850485] env[62368]: _type = "Task" [ 970.850485] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.856512] env[62368]: DEBUG nova.compute.manager [req-669e74ee-4576-4e9e-8282-0535253d0893 req-7a2c5890-8505-4675-abe6-ab099e53f9f7 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Received event network-changed-d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.856804] env[62368]: DEBUG nova.compute.manager [req-669e74ee-4576-4e9e-8282-0535253d0893 req-7a2c5890-8505-4675-abe6-ab099e53f9f7 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Refreshing instance network info cache due to event network-changed-d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 970.857367] env[62368]: DEBUG oslo_concurrency.lockutils [req-669e74ee-4576-4e9e-8282-0535253d0893 req-7a2c5890-8505-4675-abe6-ab099e53f9f7 service nova] Acquiring lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.857528] env[62368]: DEBUG oslo_concurrency.lockutils [req-669e74ee-4576-4e9e-8282-0535253d0893 req-7a2c5890-8505-4675-abe6-ab099e53f9f7 service nova] Acquired lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.857704] env[62368]: DEBUG nova.network.neutron [req-669e74ee-4576-4e9e-8282-0535253d0893 req-7a2c5890-8505-4675-abe6-ab099e53f9f7 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Refreshing network info cache for port d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.867083] env[62368]: DEBUG oslo_vmware.api [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198910, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.937027] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198907, 'name': Rename_Task, 'duration_secs': 0.174228} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.937542] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 970.937947] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-265a711e-1109-4c45-8aa3-5e137b5af65f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.946440] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 970.946440] env[62368]: value = "task-1198911" [ 970.946440] env[62368]: _type = "Task" [ 970.946440] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.955016] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.962028] env[62368]: DEBUG nova.scheduler.client.report [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.974663] env[62368]: DEBUG oslo_vmware.api [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198908, 'name': PowerOffVM_Task, 'duration_secs': 0.274502} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.975066] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 970.975342] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 970.975668] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-972daf05-9144-4d9e-9d1f-ae6e3186b828 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.051232] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 971.051456] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 971.051710] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleting the datastore file [datastore2] e427d283-1f8f-4c7a-869c-4cd48a595470 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.052042] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1d308c0-ecd6-45f6-9d3a-9c824122bfca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.060319] env[62368]: DEBUG oslo_vmware.api [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 971.060319] env[62368]: value = "task-1198913" [ 971.060319] env[62368]: _type = "Task" [ 971.060319] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.074029] env[62368]: DEBUG oslo_vmware.api [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198913, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.268847] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.293805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.308254] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198909, 'name': Rename_Task, 'duration_secs': 0.191907} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.308601] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 971.308881] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aad2a9f8-b12d-42a8-b375-d5b40cf94c2b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.315930] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 971.315930] env[62368]: value = "task-1198914" [ 971.315930] env[62368]: _type = "Task" [ 971.315930] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.323665] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.362186] env[62368]: DEBUG oslo_vmware.api [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198910, 'name': ReconfigVM_Task, 'duration_secs': 0.391429} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.362466] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfigured VM instance instance-00000058 to attach disk [datastore1] volume-cf7750c2-86cd-43ff-89ac-880138be3dc5/volume-cf7750c2-86cd-43ff-89ac-880138be3dc5.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.367893] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87e20f3b-9d22-4083-b3aa-6e063ac4a393 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.384708] env[62368]: DEBUG oslo_vmware.api [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 971.384708] env[62368]: value = "task-1198915" [ 971.384708] env[62368]: _type = "Task" [ 971.384708] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.394309] env[62368]: DEBUG oslo_vmware.api [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198915, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.457438] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198911, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.469634] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.470255] env[62368]: DEBUG nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 971.575805] env[62368]: DEBUG oslo_vmware.api [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198913, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18225} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.580310] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.580671] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 971.580985] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 971.581303] env[62368]: INFO nova.compute.manager [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Took 1.15 seconds to destroy the instance on the hypervisor. [ 971.581707] env[62368]: DEBUG oslo.service.loopingcall [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.582559] env[62368]: DEBUG nova.compute.manager [-] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.582723] env[62368]: DEBUG nova.network.neutron [-] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 971.770072] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.778373] env[62368]: DEBUG nova.network.neutron [req-669e74ee-4576-4e9e-8282-0535253d0893 req-7a2c5890-8505-4675-abe6-ab099e53f9f7 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Updated VIF entry in instance network info cache for port d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 971.778728] env[62368]: DEBUG nova.network.neutron [req-669e74ee-4576-4e9e-8282-0535253d0893 req-7a2c5890-8505-4675-abe6-ab099e53f9f7 service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Updating instance_info_cache with network_info: [{"id": "d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c", "address": "fa:16:3e:2e:88:d0", "network": {"id": "4e121ffc-3515-4d77-8958-0cd1e568ebfa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1614832061-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46f27452524540b596fa32b9999dc208", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4014180-3c", "ovs_interfaceid": "d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.825134] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813922c9-fcc5-4dfe-a992-9ca477bd1db6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.849285] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8d05fe-548d-4793-bd18-05fde48555cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.852255] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198914, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.857878] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance '72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2' progress to 83 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 971.896394] env[62368]: DEBUG oslo_vmware.api [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198915, 'name': ReconfigVM_Task, 'duration_secs': 0.168804} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.896657] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259876', 'volume_id': 'cf7750c2-86cd-43ff-89ac-880138be3dc5', 'name': 'volume-cf7750c2-86cd-43ff-89ac-880138be3dc5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e09be20-ce79-4bfa-bf84-2877f4e534e6', 'attached_at': '', 'detached_at': '', 'volume_id': 'cf7750c2-86cd-43ff-89ac-880138be3dc5', 'serial': 'cf7750c2-86cd-43ff-89ac-880138be3dc5'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 971.936893] env[62368]: DEBUG nova.compute.manager [req-7a3f4c43-57af-4851-8e02-6dac3ae633a9 req-52e18f84-5b04-46bb-b81b-3af6c5722062 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Received event network-vif-deleted-318e6d00-8a0d-4405-b405-50d208bc29f8 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.937357] env[62368]: INFO nova.compute.manager [req-7a3f4c43-57af-4851-8e02-6dac3ae633a9 req-52e18f84-5b04-46bb-b81b-3af6c5722062 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Neutron deleted interface 318e6d00-8a0d-4405-b405-50d208bc29f8; detaching it from the instance and deleting it from the info cache [ 971.937704] env[62368]: DEBUG nova.network.neutron [req-7a3f4c43-57af-4851-8e02-6dac3ae633a9 req-52e18f84-5b04-46bb-b81b-3af6c5722062 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.963042] env[62368]: DEBUG oslo_vmware.api [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198911, 'name': PowerOnVM_Task, 'duration_secs': 0.645175} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.963432] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 971.975670] env[62368]: DEBUG nova.compute.utils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 971.977476] env[62368]: DEBUG nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 971.977720] env[62368]: DEBUG nova.network.neutron [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 972.018686] env[62368]: DEBUG nova.policy [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3fbd607d8b74e22987a89cd0da5f065', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e31b8f1352574bb7808b06e732da7e4e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 972.071639] env[62368]: DEBUG nova.compute.manager [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.072593] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ec0cd0-cece-45fe-a269-17d7a1b4e13a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.269773] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.281464] env[62368]: DEBUG oslo_concurrency.lockutils [req-669e74ee-4576-4e9e-8282-0535253d0893 req-7a2c5890-8505-4675-abe6-ab099e53f9f7 service nova] Releasing lock "refresh_cache-c1b238e7-3636-47db-b86e-3fcfe917f705" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.330841] env[62368]: DEBUG oslo_vmware.api [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198914, 'name': PowerOnVM_Task, 'duration_secs': 0.568199} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.332062] env[62368]: DEBUG nova.network.neutron [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Successfully created port: 76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 972.334084] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 972.334360] env[62368]: INFO nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Took 9.70 seconds to spawn the instance on the hypervisor. [ 972.334646] env[62368]: DEBUG nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.335531] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa803970-086e-4c15-9e3d-9b6b7e079ce7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.364114] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 972.368454] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a850932-f606-4277-835a-6c52c7d58921 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.375830] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 972.375830] env[62368]: value = "task-1198916" [ 972.375830] env[62368]: _type = "Task" [ 972.375830] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.385969] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.403035] env[62368]: DEBUG nova.network.neutron [-] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.444198] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-988faecf-0c3a-4c51-a253-df324a35a925 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.454033] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cba05a-2f73-4e8d-8fa1-aa097ada80fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.490099] env[62368]: DEBUG nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 972.493578] env[62368]: DEBUG nova.compute.manager [req-7a3f4c43-57af-4851-8e02-6dac3ae633a9 req-52e18f84-5b04-46bb-b81b-3af6c5722062 service nova] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Detach interface failed, port_id=318e6d00-8a0d-4405-b405-50d208bc29f8, reason: Instance e427d283-1f8f-4c7a-869c-4cd48a595470 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 972.590572] env[62368]: DEBUG oslo_concurrency.lockutils [None req-732df7df-39b2-4e54-a6da-d32679647046 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.248s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.770591] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.855560] env[62368]: INFO nova.compute.manager [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Took 20.06 seconds to build instance. [ 972.889056] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198916, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.905804] env[62368]: INFO nova.compute.manager [-] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Took 1.32 seconds to deallocate network for instance. [ 972.949340] env[62368]: DEBUG nova.objects.instance [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lazy-loading 'flavor' on Instance uuid 3e09be20-ce79-4bfa-bf84-2877f4e534e6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.271182] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.358205] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7bcad769-a91e-4475-8c8c-42458da4555e tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.205s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.373029] env[62368]: DEBUG oslo_concurrency.lockutils [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "84bb7208-670f-4e1d-862a-a480f7db0f42" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.373325] env[62368]: DEBUG oslo_concurrency.lockutils [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.373542] env[62368]: DEBUG nova.compute.manager [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.374499] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2818c23a-d52f-4e41-81f5-565729789646 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.386140] env[62368]: DEBUG oslo_vmware.api [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198916, 'name': PowerOnVM_Task, 'duration_secs': 0.924338} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.387783] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 973.387985] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ea58a40a-24dc-4406-9bb3-ab49d2f66653 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance '72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2' progress to 100 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 973.391631] env[62368]: DEBUG nova.compute.manager [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62368) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 973.392202] env[62368]: DEBUG nova.objects.instance [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'flavor' on Instance uuid 84bb7208-670f-4e1d-862a-a480f7db0f42 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.413237] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.413704] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.413902] env[62368]: DEBUG nova.objects.instance [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lazy-loading 'resources' on Instance uuid e427d283-1f8f-4c7a-869c-4cd48a595470 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.454463] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d57ade7-1270-45a8-8a45-1b165f11bff8 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 11.343s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.501068] env[62368]: DEBUG nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 973.530297] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 973.530560] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 973.530722] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.530912] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 973.531072] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.531226] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 973.531433] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 973.531595] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 973.531764] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 973.531931] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 973.532126] env[62368]: DEBUG nova.virt.hardware [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 973.533309] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9b0dd8-a52a-48f0-96d3-636f3b6fbdf1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.542220] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2287c8f-cc9e-4268-9b29-763bbed294dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.774346] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.796203] env[62368]: INFO nova.compute.manager [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Rescuing [ 973.796497] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.796678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.796885] env[62368]: DEBUG nova.network.neutron [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 973.815705] env[62368]: DEBUG nova.compute.manager [req-1ab40834-c74b-4726-9b39-0491c27171bb req-288bc5e8-2c1f-476a-8d8f-13f12f79637a service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Received event network-vif-plugged-76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.815885] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ab40834-c74b-4726-9b39-0491c27171bb req-288bc5e8-2c1f-476a-8d8f-13f12f79637a service nova] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.816150] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ab40834-c74b-4726-9b39-0491c27171bb req-288bc5e8-2c1f-476a-8d8f-13f12f79637a service nova] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.816330] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ab40834-c74b-4726-9b39-0491c27171bb req-288bc5e8-2c1f-476a-8d8f-13f12f79637a service nova] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.816509] env[62368]: DEBUG nova.compute.manager [req-1ab40834-c74b-4726-9b39-0491c27171bb req-288bc5e8-2c1f-476a-8d8f-13f12f79637a service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] No waiting events found dispatching network-vif-plugged-76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 973.816693] env[62368]: WARNING nova.compute.manager [req-1ab40834-c74b-4726-9b39-0491c27171bb req-288bc5e8-2c1f-476a-8d8f-13f12f79637a service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Received unexpected event network-vif-plugged-76aee038-386b-401a-8c49-5ee204436ab3 for instance with vm_state building and task_state spawning. [ 973.898984] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 973.899273] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f1d2817-4915-43cd-a04c-c4d06c6bd57e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.907379] env[62368]: DEBUG oslo_vmware.api [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 973.907379] env[62368]: value = "task-1198917" [ 973.907379] env[62368]: _type = "Task" [ 973.907379] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.919036] env[62368]: DEBUG oslo_vmware.api [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.162140] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a601fc8-2f7d-463f-958d-7fc11958d449 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.175114] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6cec5b-ac48-418d-b63a-b4316b8ddcf7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.209909] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6151ce9-ca1f-4338-8618-7062165ff764 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.218773] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2941b5-973c-4936-b4bf-e024fe543594 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.233108] env[62368]: DEBUG nova.compute.provider_tree [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.272945] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.363704] env[62368]: DEBUG nova.network.neutron [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Successfully updated port: 76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 974.418464] env[62368]: DEBUG oslo_vmware.api [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198917, 'name': PowerOffVM_Task, 'duration_secs': 0.240355} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.418750] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 974.418929] env[62368]: DEBUG nova.compute.manager [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.419727] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8de6447-5f87-4199-b7f7-a49d18e0a651 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.432623] env[62368]: DEBUG nova.compute.manager [req-41c43e8c-f41f-4e54-b51d-4a7103781718 req-0e291bc2-2505-44aa-9649-a40d548e2f2a service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Received event network-changed-76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.432823] env[62368]: DEBUG nova.compute.manager [req-41c43e8c-f41f-4e54-b51d-4a7103781718 req-0e291bc2-2505-44aa-9649-a40d548e2f2a service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Refreshing instance network info cache due to event network-changed-76aee038-386b-401a-8c49-5ee204436ab3. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 974.433068] env[62368]: DEBUG oslo_concurrency.lockutils [req-41c43e8c-f41f-4e54-b51d-4a7103781718 req-0e291bc2-2505-44aa-9649-a40d548e2f2a service nova] Acquiring lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.433224] env[62368]: DEBUG oslo_concurrency.lockutils [req-41c43e8c-f41f-4e54-b51d-4a7103781718 req-0e291bc2-2505-44aa-9649-a40d548e2f2a service nova] Acquired lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.433532] env[62368]: DEBUG nova.network.neutron [req-41c43e8c-f41f-4e54-b51d-4a7103781718 req-0e291bc2-2505-44aa-9649-a40d548e2f2a service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Refreshing network info cache for port 76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 974.517802] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95757c9c-736e-4eef-b907-6d9d4347a75e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.525112] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-845ce53e-a2b9-43a1-b9ee-0df02f636957 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Suspending the VM {{(pid=62368) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 974.525396] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ec9100d7-7b7c-460e-a647-6742a12283bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.536669] env[62368]: DEBUG oslo_vmware.api [None req-845ce53e-a2b9-43a1-b9ee-0df02f636957 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 974.536669] env[62368]: value = "task-1198918" [ 974.536669] env[62368]: _type = "Task" [ 974.536669] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.545310] env[62368]: DEBUG oslo_vmware.api [None req-845ce53e-a2b9-43a1-b9ee-0df02f636957 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198918, 'name': SuspendVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.736040] env[62368]: DEBUG nova.scheduler.client.report [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.773492] env[62368]: DEBUG oslo_vmware.api [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198903, 'name': ReconfigVM_Task, 'duration_secs': 5.784566} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.773743] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.774358] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Reconfigured VM to detach interface {{(pid=62368) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 974.866936] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.879845] env[62368]: DEBUG nova.network.neutron [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating instance_info_cache with network_info: [{"id": "561cb794-4c95-416d-825a-193df8ddc3ca", "address": "fa:16:3e:2b:4e:03", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap561cb794-4c", "ovs_interfaceid": "561cb794-4c95-416d-825a-193df8ddc3ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.942243] env[62368]: DEBUG oslo_concurrency.lockutils [None req-70ebdfa1-5ff8-47b7-ae64-47e353f83eab tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.569s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.988354] env[62368]: DEBUG nova.network.neutron [req-41c43e8c-f41f-4e54-b51d-4a7103781718 req-0e291bc2-2505-44aa-9649-a40d548e2f2a service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 975.050246] env[62368]: DEBUG oslo_vmware.api [None req-845ce53e-a2b9-43a1-b9ee-0df02f636957 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198918, 'name': SuspendVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.116836] env[62368]: DEBUG nova.network.neutron [req-41c43e8c-f41f-4e54-b51d-4a7103781718 req-0e291bc2-2505-44aa-9649-a40d548e2f2a service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.245080] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.283130] env[62368]: INFO nova.scheduler.client.report [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted allocations for instance e427d283-1f8f-4c7a-869c-4cd48a595470 [ 975.383343] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.547962] env[62368]: DEBUG oslo_vmware.api [None req-845ce53e-a2b9-43a1-b9ee-0df02f636957 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198918, 'name': SuspendVM_Task, 'duration_secs': 0.734522} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.548288] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-845ce53e-a2b9-43a1-b9ee-0df02f636957 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Suspended the VM {{(pid=62368) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 975.548481] env[62368]: DEBUG nova.compute.manager [None req-845ce53e-a2b9-43a1-b9ee-0df02f636957 tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.549272] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8178d86c-2326-46d8-9fa1-abd9272ac10e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.567559] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.567838] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.568076] env[62368]: DEBUG nova.compute.manager [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Going to confirm migration 4 {{(pid=62368) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 975.619915] env[62368]: DEBUG oslo_concurrency.lockutils [req-41c43e8c-f41f-4e54-b51d-4a7103781718 req-0e291bc2-2505-44aa-9649-a40d548e2f2a service nova] Releasing lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.620450] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.620721] env[62368]: DEBUG nova.network.neutron [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 975.791749] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03da4443-b655-4204-a0ac-08518b8434dc tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "e427d283-1f8f-4c7a-869c-4cd48a595470" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.378s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.839795] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "84bb7208-670f-4e1d-862a-a480f7db0f42" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.840227] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.840361] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "84bb7208-670f-4e1d-862a-a480f7db0f42-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.840566] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.840745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.842736] env[62368]: INFO nova.compute.manager [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Terminating instance [ 975.844453] env[62368]: DEBUG nova.compute.manager [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.844852] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 975.845490] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0308bfcd-4ef1-4a86-b19f-6761416955aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.854563] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 975.855174] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a65474e7-43a2-4480-82e5-9e1740bb7512 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.918633] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 975.918917] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98e25a07-9d2e-4de7-8aeb-867aeef79a87 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.926859] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 975.926859] env[62368]: value = "task-1198920" [ 975.926859] env[62368]: _type = "Task" [ 975.926859] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.937502] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.941218] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 975.941502] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 975.941755] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleting the datastore file [datastore1] 84bb7208-670f-4e1d-862a-a480f7db0f42 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.942065] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc2f2b64-1857-4116-aec6-44e3ee03b098 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.949990] env[62368]: DEBUG oslo_vmware.api [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 975.949990] env[62368]: value = "task-1198921" [ 975.949990] env[62368]: _type = "Task" [ 975.949990] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.958766] env[62368]: DEBUG oslo_vmware.api [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.148339] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.148465] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.148678] env[62368]: DEBUG nova.network.neutron [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 976.148908] env[62368]: DEBUG nova.objects.instance [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lazy-loading 'info_cache' on Instance uuid 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.163409] env[62368]: DEBUG nova.network.neutron [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 976.298924] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.298924] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquired lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.298924] env[62368]: DEBUG nova.network.neutron [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 976.378499] env[62368]: DEBUG nova.network.neutron [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Updating instance_info_cache with network_info: [{"id": "76aee038-386b-401a-8c49-5ee204436ab3", "address": "fa:16:3e:46:83:dc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76aee038-38", "ovs_interfaceid": "76aee038-386b-401a-8c49-5ee204436ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.381078] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.381078] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.437288] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198920, 'name': PowerOffVM_Task, 'duration_secs': 0.289805} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.437603] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 976.438467] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a060ca95-ff2a-432e-aa89-ac1973b2912a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.464554] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fa3fab-ba06-42f8-84e7-df39a6142e2a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.477302] env[62368]: DEBUG oslo_vmware.api [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264097} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.478037] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.478037] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 976.478037] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 976.478203] env[62368]: INFO nova.compute.manager [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Took 0.63 seconds to destroy the instance on the hypervisor. [ 976.478453] env[62368]: DEBUG oslo.service.loopingcall [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.478608] env[62368]: DEBUG nova.compute.manager [-] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.478710] env[62368]: DEBUG nova.network.neutron [-] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 976.499559] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.499885] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12060444-d6fd-4dc8-92da-489b815b5252 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.508109] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 976.508109] env[62368]: value = "task-1198922" [ 976.508109] env[62368]: _type = "Task" [ 976.508109] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.521241] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 976.521467] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.521706] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.521862] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.522053] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.522301] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f007440d-85cb-46fe-b1b7-a6727e6e448c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.532310] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.532500] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 976.533299] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb187fa9-8735-4348-888f-aa470db2b884 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.540533] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 976.540533] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526cdaaf-3dc9-9623-f432-c6cfb1fa8e05" [ 976.540533] env[62368]: _type = "Task" [ 976.540533] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.554380] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526cdaaf-3dc9-9623-f432-c6cfb1fa8e05, 'name': SearchDatastore_Task, 'duration_secs': 0.011855} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.555807] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67d255f8-2692-4849-b481-86b685ab0e75 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.561689] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 976.561689] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52778e60-66c2-cead-2683-4c7332af1216" [ 976.561689] env[62368]: _type = "Task" [ 976.561689] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.570750] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52778e60-66c2-cead-2683-4c7332af1216, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.619111] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "5633292c-9a74-4c2d-893b-6823f4a587a2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.619416] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.619648] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.619861] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.620059] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.622422] env[62368]: INFO nova.compute.manager [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Terminating instance [ 976.624383] env[62368]: DEBUG nova.compute.manager [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.624656] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 976.625444] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59638ac-c362-40a3-98b0-3dda2cf3c806 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.634373] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.634671] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-285d21a3-e945-4f74-8c3a-17ffb083291d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.642447] env[62368]: DEBUG oslo_vmware.api [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 976.642447] env[62368]: value = "task-1198923" [ 976.642447] env[62368]: _type = "Task" [ 976.642447] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.656428] env[62368]: DEBUG oslo_vmware.api [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198923, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.808891] env[62368]: DEBUG nova.compute.manager [req-7dad0215-3fd7-4464-9742-6a23f002d5b2 req-24ce5dc7-2188-439a-a1fc-e93d274a71b6 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Received event network-vif-deleted-114fbe1a-b6e9-405a-9d63-241a22cb6560 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.809230] env[62368]: INFO nova.compute.manager [req-7dad0215-3fd7-4464-9742-6a23f002d5b2 req-24ce5dc7-2188-439a-a1fc-e93d274a71b6 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Neutron deleted interface 114fbe1a-b6e9-405a-9d63-241a22cb6560; detaching it from the instance and deleting it from the info cache [ 976.809445] env[62368]: DEBUG nova.network.neutron [req-7dad0215-3fd7-4464-9742-6a23f002d5b2 req-24ce5dc7-2188-439a-a1fc-e93d274a71b6 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.883733] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Releasing lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.884110] env[62368]: DEBUG nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Instance network_info: |[{"id": "76aee038-386b-401a-8c49-5ee204436ab3", "address": "fa:16:3e:46:83:dc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76aee038-38", "ovs_interfaceid": "76aee038-386b-401a-8c49-5ee204436ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 976.884558] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:83:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '975b168a-03e5-449d-95ac-4d51ba027242', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76aee038-386b-401a-8c49-5ee204436ab3', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 976.892357] env[62368]: DEBUG oslo.service.loopingcall [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.892656] env[62368]: DEBUG nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 976.897949] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 976.898415] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b77caa1-ecfe-4261-bf2f-53b48f4cde3c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.920058] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.920058] env[62368]: value = "task-1198924" [ 976.920058] env[62368]: _type = "Task" [ 976.920058] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.928390] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198924, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.022619] env[62368]: INFO nova.compute.manager [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Resuming [ 977.023288] env[62368]: DEBUG nova.objects.instance [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lazy-loading 'flavor' on Instance uuid eea21546-fbbf-4440-829c-8583c4ccabb6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.073051] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52778e60-66c2-cead-2683-4c7332af1216, 'name': SearchDatastore_Task, 'duration_secs': 0.010471} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.073851] env[62368]: INFO nova.network.neutron [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Port d1b97d93-2fbd-4ff2-948f-6f197d6ed77f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 977.074178] env[62368]: DEBUG nova.network.neutron [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [{"id": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "address": "fa:16:3e:95:85:5e", "network": {"id": "3ee164de-30af-4404-935a-a7f988d824b1", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-940834407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcecae6c09124e9fa4dcd5d63833b0bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14840bf2-7d", "ovs_interfaceid": "14840bf2-7da7-4d91-aa96-02a1771fb7d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.075323] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.076067] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. {{(pid=62368) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 977.076165] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7fd2e04-3df7-43f0-8e89-4b340a627ca2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.083616] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 977.083616] env[62368]: value = "task-1198925" [ 977.083616] env[62368]: _type = "Task" [ 977.083616] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.092380] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198925, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.154419] env[62368]: DEBUG oslo_vmware.api [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198923, 'name': PowerOffVM_Task, 'duration_secs': 0.177045} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.154748] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.154947] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 977.155290] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f02298a1-013a-4b18-9fe0-87c8778db908 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.239994] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 977.240414] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 977.240742] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleting the datastore file [datastore1] 5633292c-9a74-4c2d-893b-6823f4a587a2 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.241169] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cffcdae0-e76d-4e09-864b-20a4e4453bf2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.249341] env[62368]: DEBUG oslo_vmware.api [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 977.249341] env[62368]: value = "task-1198927" [ 977.249341] env[62368]: _type = "Task" [ 977.249341] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.267120] env[62368]: DEBUG oslo_vmware.api [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.279391] env[62368]: DEBUG nova.network.neutron [-] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.312359] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc999883-6283-40bc-8480-1294bef0bc29 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.324155] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b628ac4-d509-44c4-bccb-fbba7fd1e2ce {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.365376] env[62368]: DEBUG nova.compute.manager [req-7dad0215-3fd7-4464-9742-6a23f002d5b2 req-24ce5dc7-2188-439a-a1fc-e93d274a71b6 service nova] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Detach interface failed, port_id=114fbe1a-b6e9-405a-9d63-241a22cb6560, reason: Instance 84bb7208-670f-4e1d-862a-a480f7db0f42 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 977.430136] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.430136] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.430556] env[62368]: INFO nova.compute.claims [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.442677] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198924, 'name': CreateVM_Task, 'duration_secs': 0.441868} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.443152] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 977.445042] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.445129] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.445557] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.445922] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d22111d-3d8f-4f45-8e11-d4715419531d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.452366] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 977.452366] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52537c8a-9e8b-655a-32b6-c93cde30587a" [ 977.452366] env[62368]: _type = "Task" [ 977.452366] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.462886] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52537c8a-9e8b-655a-32b6-c93cde30587a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.484566] env[62368]: DEBUG nova.network.neutron [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance_info_cache with network_info: [{"id": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "address": "fa:16:3e:b0:23:0b", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd229b53b-0e", "ovs_interfaceid": "d229b53b-0e94-41c9-b8e9-0b0eaa366d3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.579053] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Releasing lock "refresh_cache-5633292c-9a74-4c2d-893b-6823f4a587a2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.596073] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198925, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.760260] env[62368]: DEBUG oslo_vmware.api [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.416801} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.760443] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 977.760635] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 977.760820] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 977.761013] env[62368]: INFO nova.compute.manager [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Took 1.14 seconds to destroy the instance on the hypervisor. [ 977.761310] env[62368]: DEBUG oslo.service.loopingcall [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.761515] env[62368]: DEBUG nova.compute.manager [-] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 977.761613] env[62368]: DEBUG nova.network.neutron [-] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 977.783351] env[62368]: INFO nova.compute.manager [-] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Took 1.30 seconds to deallocate network for instance. [ 977.966225] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52537c8a-9e8b-655a-32b6-c93cde30587a, 'name': SearchDatastore_Task, 'duration_secs': 0.051952} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.966478] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.966780] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.966991] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.967184] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.967374] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.967645] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b78e163-707c-4c17-9e9f-b30bbf246811 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.976585] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.976783] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 977.977502] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74c3bad4-a328-409f-90b6-1cc5924ca5ff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.983173] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 977.983173] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52092590-0864-f3ed-787e-3f15186fe0de" [ 977.983173] env[62368]: _type = "Task" [ 977.983173] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.991640] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.991875] env[62368]: DEBUG nova.objects.instance [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lazy-loading 'migration_context' on Instance uuid 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.993283] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52092590-0864-f3ed-787e-3f15186fe0de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.032588] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.032768] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquired lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.033338] env[62368]: DEBUG nova.network.neutron [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 978.082312] env[62368]: DEBUG oslo_concurrency.lockutils [None req-83ed7e5f-7821-4fc7-b680-6465b0e0d219 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "interface-5633292c-9a74-4c2d-893b-6823f4a587a2-d1b97d93-2fbd-4ff2-948f-6f197d6ed77f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.903s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.095866] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198925, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.593473} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.096398] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. [ 978.097849] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2873a914-f1da-4a61-9a32-3ab3998487b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.130321] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.130997] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b014a5bc-6662-4528-9172-9504faf003ef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.155228] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 978.155228] env[62368]: value = "task-1198928" [ 978.155228] env[62368]: _type = "Task" [ 978.155228] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.292255] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.495977] env[62368]: DEBUG nova.objects.base [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Object Instance<72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2> lazy-loaded attributes: info_cache,migration_context {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 978.496350] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52092590-0864-f3ed-787e-3f15186fe0de, 'name': SearchDatastore_Task, 'duration_secs': 0.009673} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.498392] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c284e19-ece7-45d9-9185-ce2925819f81 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.501753] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8bba44f-6819-400c-a476-6329008bef8a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.533252] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c56c716-a72c-4b75-9764-7fbbfbec5771 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.535654] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 978.535654] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5227ea2f-310b-cc67-8930-8cd70b998bfc" [ 978.535654] env[62368]: _type = "Task" [ 978.535654] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.544683] env[62368]: DEBUG oslo_vmware.api [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 978.544683] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52743b85-a5a9-f703-a589-38c29e783593" [ 978.544683] env[62368]: _type = "Task" [ 978.544683] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.556948] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5227ea2f-310b-cc67-8930-8cd70b998bfc, 'name': SearchDatastore_Task, 'duration_secs': 0.010386} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.558119] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.558290] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] ca1c6fcb-9450-4b59-b8b1-39e88aa61af0/ca1c6fcb-9450-4b59-b8b1-39e88aa61af0.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 978.558632] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fd472fd-7384-4e97-a579-5ba30e72cd9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.565084] env[62368]: DEBUG oslo_vmware.api [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52743b85-a5a9-f703-a589-38c29e783593, 'name': SearchDatastore_Task, 'duration_secs': 0.007205} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.565734] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.571070] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 978.571070] env[62368]: value = "task-1198929" [ 978.571070] env[62368]: _type = "Task" [ 978.571070] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.580521] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198929, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.667884] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198928, 'name': ReconfigVM_Task, 'duration_secs': 0.403146} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.668381] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.669461] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57fbe6a-bdb4-43ea-8dde-e52df2588d25 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.715092] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b63d7857-f375-4d41-8af4-a083c9ff4f94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.732914] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 978.732914] env[62368]: value = "task-1198930" [ 978.732914] env[62368]: _type = "Task" [ 978.732914] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.745145] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198930, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.767497] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570b8dde-f639-48ba-bad1-4b9f3d2c470e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.775008] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9feb9a20-756f-45d8-a71c-1131429690f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.807583] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727e7aa7-471a-4ba6-8ae4-a7afa8ec6bac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.815040] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fceabe8-1f47-48bc-bcea-6aea01b7b369 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.829910] env[62368]: DEBUG nova.compute.provider_tree [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.834061] env[62368]: DEBUG nova.compute.manager [req-f21a3b06-ddf8-4c24-a53d-d243feca9c87 req-046503ff-fd8c-4f80-9a0e-8beb55d62633 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Received event network-vif-deleted-14840bf2-7da7-4d91-aa96-02a1771fb7d1 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.834261] env[62368]: INFO nova.compute.manager [req-f21a3b06-ddf8-4c24-a53d-d243feca9c87 req-046503ff-fd8c-4f80-9a0e-8beb55d62633 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Neutron deleted interface 14840bf2-7da7-4d91-aa96-02a1771fb7d1; detaching it from the instance and deleting it from the info cache [ 978.834530] env[62368]: DEBUG nova.network.neutron [req-f21a3b06-ddf8-4c24-a53d-d243feca9c87 req-046503ff-fd8c-4f80-9a0e-8beb55d62633 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.865247] env[62368]: DEBUG nova.network.neutron [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating instance_info_cache with network_info: [{"id": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "address": "fa:16:3e:d1:c4:2e", "network": {"id": "ff955407-6db9-436d-ac96-9919ab1eec96", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1959365261-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e250606f3df240eb93ce45925b00ad43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ebe893-6b", "ovs_interfaceid": "b2ebe893-6ba3-40da-8def-fb6cf1ab442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.002983] env[62368]: DEBUG nova.network.neutron [-] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.081571] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198929, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.243502] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198930, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.334389] env[62368]: DEBUG nova.scheduler.client.report [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.337258] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b66120c4-0f83-4ced-944a-8213961b04de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.347778] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0b7fe9-d7a8-49d0-8de4-67a7e54353f4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.367145] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Releasing lock "refresh_cache-eea21546-fbbf-4440-829c-8583c4ccabb6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.367956] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831feb3f-64ba-4813-847c-def75bf67fcc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.380753] env[62368]: DEBUG nova.compute.manager [req-f21a3b06-ddf8-4c24-a53d-d243feca9c87 req-046503ff-fd8c-4f80-9a0e-8beb55d62633 service nova] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Detach interface failed, port_id=14840bf2-7da7-4d91-aa96-02a1771fb7d1, reason: Instance 5633292c-9a74-4c2d-893b-6823f4a587a2 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 979.384781] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Resuming the VM {{(pid=62368) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 979.385072] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4593e8ad-a0de-410e-af18-e07098c0b242 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.391654] env[62368]: DEBUG oslo_vmware.api [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 979.391654] env[62368]: value = "task-1198931" [ 979.391654] env[62368]: _type = "Task" [ 979.391654] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.405864] env[62368]: DEBUG oslo_vmware.api [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198931, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.505885] env[62368]: INFO nova.compute.manager [-] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Took 1.74 seconds to deallocate network for instance. [ 979.584664] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198929, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.744565] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198930, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.838762] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.839496] env[62368]: DEBUG nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 979.842488] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.550s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.842862] env[62368]: DEBUG nova.objects.instance [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'resources' on Instance uuid 84bb7208-670f-4e1d-862a-a480f7db0f42 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.901428] env[62368]: DEBUG oslo_vmware.api [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198931, 'name': PowerOnVM_Task} progress is 93%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.013422] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.086747] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198929, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.299361} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.087034] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] ca1c6fcb-9450-4b59-b8b1-39e88aa61af0/ca1c6fcb-9450-4b59-b8b1-39e88aa61af0.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 980.087300] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 980.087578] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1a394f1-336d-437e-8b49-cf2b35a06011 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.099325] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 980.099325] env[62368]: value = "task-1198932" [ 980.099325] env[62368]: _type = "Task" [ 980.099325] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.111539] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198932, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.244252] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198930, 'name': ReconfigVM_Task, 'duration_secs': 1.040204} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.244606] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 980.244876] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca1d9746-b75f-4212-9dd4-c73a77643f4b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.252267] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 980.252267] env[62368]: value = "task-1198933" [ 980.252267] env[62368]: _type = "Task" [ 980.252267] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.260392] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.346220] env[62368]: DEBUG nova.compute.utils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 980.347631] env[62368]: DEBUG nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 980.347803] env[62368]: DEBUG nova.network.neutron [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 980.400441] env[62368]: DEBUG nova.policy [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c66c425264424e3eaa37c8058e62cb30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '026a3176abc34001ab3506445226777f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.408518] env[62368]: DEBUG oslo_vmware.api [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198931, 'name': PowerOnVM_Task, 'duration_secs': 0.776731} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.408821] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Resumed the VM {{(pid=62368) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 980.409124] env[62368]: DEBUG nova.compute.manager [None req-ae9f5654-4881-46b6-9b95-550e9d2d628d tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.410311] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35db744c-3911-4d78-8857-dafe7c18ce6c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.574525] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c88791a-a002-41a1-8d9e-78d5b42efebd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.582707] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1207dc-4247-4de7-ab97-2a89060924ab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.620297] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fd3e08-916f-49d0-837c-bbc9e2001752 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.632925] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274c4e29-d6c5-4ccb-ae1c-501bac689443 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.637801] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198932, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.130781} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.638174] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.639466] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01eeaf1-c08d-4218-9f61-dcb854204a3c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.651369] env[62368]: DEBUG nova.compute.provider_tree [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.674012] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] ca1c6fcb-9450-4b59-b8b1-39e88aa61af0/ca1c6fcb-9450-4b59-b8b1-39e88aa61af0.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.676558] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c77a3aee-6a5e-4411-abcf-1bb4ddced38d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.690379] env[62368]: DEBUG nova.network.neutron [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Successfully created port: b78fda7d-cbff-4736-9fcf-49ad36e29f67 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.698540] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 980.698540] env[62368]: value = "task-1198934" [ 980.698540] env[62368]: _type = "Task" [ 980.698540] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.706868] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198934, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.763012] env[62368]: DEBUG oslo_vmware.api [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198933, 'name': PowerOnVM_Task, 'duration_secs': 0.43192} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.763309] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 980.766112] env[62368]: DEBUG nova.compute.manager [None req-fc2c6510-7106-43e9-9d04-e3d05aabcf98 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.766907] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034690ec-e548-40dd-bbee-3410d5202e49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.852763] env[62368]: DEBUG nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.173882] env[62368]: ERROR nova.scheduler.client.report [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [req-4100bf6c-0778-41ed-b6c9-7cb541752d62] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4100bf6c-0778-41ed-b6c9-7cb541752d62"}]} [ 981.191460] env[62368]: DEBUG nova.scheduler.client.report [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 981.213563] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198934, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.214772] env[62368]: DEBUG nova.scheduler.client.report [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 981.215025] env[62368]: DEBUG nova.compute.provider_tree [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.226179] env[62368]: DEBUG nova.scheduler.client.report [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 981.245656] env[62368]: DEBUG nova.scheduler.client.report [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 981.482571] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb65ac49-804f-4342-bb1d-2d1cedd07d9e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.491495] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9d6558-543c-4b88-a056-8e71d29a30f8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.526875] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10a95cd-764a-4884-9ffd-a375643e2c92 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.535336] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13df95b-b365-44b6-8ba9-8309777a55ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.549346] env[62368]: DEBUG nova.compute.provider_tree [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.708960] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198934, 'name': ReconfigVM_Task, 'duration_secs': 0.814511} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.709381] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfigured VM instance instance-00000063 to attach disk [datastore2] ca1c6fcb-9450-4b59-b8b1-39e88aa61af0/ca1c6fcb-9450-4b59-b8b1-39e88aa61af0.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.710100] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e365e9ea-e910-44b3-bcb5-1dfe90b8ff37 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.717496] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 981.717496] env[62368]: value = "task-1198935" [ 981.717496] env[62368]: _type = "Task" [ 981.717496] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.725984] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198935, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.863093] env[62368]: DEBUG nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 981.888584] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.888840] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.889048] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.889248] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.889399] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.889547] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.889785] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.889976] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.890168] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.890332] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.890648] env[62368]: DEBUG nova.virt.hardware [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.891692] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d184f67c-59d3-4f54-ad66-73ffa01e471b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.900434] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e95d5a4-983b-4ffa-af94-b6f22effe51c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.083608] env[62368]: DEBUG nova.scheduler.client.report [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 125 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 982.083934] env[62368]: DEBUG nova.compute.provider_tree [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 125 to 126 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 982.084148] env[62368]: DEBUG nova.compute.provider_tree [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.228839] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198935, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.322270] env[62368]: INFO nova.compute.manager [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Unrescuing [ 982.322534] env[62368]: DEBUG oslo_concurrency.lockutils [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.322689] env[62368]: DEBUG oslo_concurrency.lockutils [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquired lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.322857] env[62368]: DEBUG nova.network.neutron [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 982.326130] env[62368]: DEBUG nova.compute.manager [req-6de7be7e-71da-4ea0-9c1d-258df0b8cadb req-2e53cbf2-4383-4c0d-82be-a7dd69225fc3 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Received event network-vif-plugged-b78fda7d-cbff-4736-9fcf-49ad36e29f67 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.326130] env[62368]: DEBUG oslo_concurrency.lockutils [req-6de7be7e-71da-4ea0-9c1d-258df0b8cadb req-2e53cbf2-4383-4c0d-82be-a7dd69225fc3 service nova] Acquiring lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.326130] env[62368]: DEBUG oslo_concurrency.lockutils [req-6de7be7e-71da-4ea0-9c1d-258df0b8cadb req-2e53cbf2-4383-4c0d-82be-a7dd69225fc3 service nova] Lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.326130] env[62368]: DEBUG oslo_concurrency.lockutils [req-6de7be7e-71da-4ea0-9c1d-258df0b8cadb req-2e53cbf2-4383-4c0d-82be-a7dd69225fc3 service nova] Lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.326130] env[62368]: DEBUG nova.compute.manager [req-6de7be7e-71da-4ea0-9c1d-258df0b8cadb req-2e53cbf2-4383-4c0d-82be-a7dd69225fc3 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] No waiting events found dispatching network-vif-plugged-b78fda7d-cbff-4736-9fcf-49ad36e29f67 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 982.326130] env[62368]: WARNING nova.compute.manager [req-6de7be7e-71da-4ea0-9c1d-258df0b8cadb req-2e53cbf2-4383-4c0d-82be-a7dd69225fc3 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Received unexpected event network-vif-plugged-b78fda7d-cbff-4736-9fcf-49ad36e29f67 for instance with vm_state building and task_state spawning. [ 982.365993] env[62368]: DEBUG nova.network.neutron [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Successfully updated port: b78fda7d-cbff-4736-9fcf-49ad36e29f67 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 982.590033] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.747s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.592113] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.026s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.611396] env[62368]: INFO nova.scheduler.client.report [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted allocations for instance 84bb7208-670f-4e1d-862a-a480f7db0f42 [ 982.729321] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198935, 'name': Rename_Task, 'duration_secs': 0.882647} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.729690] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 982.729847] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42ac2548-bc26-4a09-a348-3991e9456944 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.737429] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 982.737429] env[62368]: value = "task-1198936" [ 982.737429] env[62368]: _type = "Task" [ 982.737429] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.756763] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198936, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.873526] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "refresh_cache-66e96455-fbe8-4a0d-8fe2-9ad60d5abade" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.873526] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "refresh_cache-66e96455-fbe8-4a0d-8fe2-9ad60d5abade" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.873526] env[62368]: DEBUG nova.network.neutron [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 983.119181] env[62368]: DEBUG oslo_concurrency.lockutils [None req-632e7592-8a6f-4144-b267-c9fce2851829 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "84bb7208-670f-4e1d-862a-a480f7db0f42" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.279s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.138620] env[62368]: DEBUG nova.network.neutron [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating instance_info_cache with network_info: [{"id": "561cb794-4c95-416d-825a-193df8ddc3ca", "address": "fa:16:3e:2b:4e:03", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap561cb794-4c", "ovs_interfaceid": "561cb794-4c95-416d-825a-193df8ddc3ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.249140] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198936, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.331784] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19e73f7-5999-4a87-b064-1d271b575cbf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.340397] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80fae02-6f7f-44d0-bc2d-1f8ad8d74295 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.376526] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870ac15c-e64c-4267-a220-d7ba51b0f872 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.387983] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eceb044-7cd9-41a9-9407-50f9d49c3cec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.402785] env[62368]: DEBUG nova.compute.provider_tree [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.424366] env[62368]: DEBUG nova.network.neutron [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 983.579352] env[62368]: DEBUG nova.network.neutron [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Updating instance_info_cache with network_info: [{"id": "b78fda7d-cbff-4736-9fcf-49ad36e29f67", "address": "fa:16:3e:c0:9d:3c", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78fda7d-cb", "ovs_interfaceid": "b78fda7d-cbff-4736-9fcf-49ad36e29f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.641777] env[62368]: DEBUG oslo_concurrency.lockutils [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Releasing lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.642607] env[62368]: DEBUG nova.objects.instance [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lazy-loading 'flavor' on Instance uuid 3e09be20-ce79-4bfa-bf84-2877f4e534e6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.750074] env[62368]: DEBUG oslo_vmware.api [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1198936, 'name': PowerOnVM_Task, 'duration_secs': 0.529507} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.750074] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 983.750074] env[62368]: INFO nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Took 10.25 seconds to spawn the instance on the hypervisor. [ 983.750074] env[62368]: DEBUG nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.750551] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27883d6c-dad3-41a2-8e4a-caa4526e0700 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.905632] env[62368]: DEBUG nova.scheduler.client.report [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.016048] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.016227] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.082216] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "refresh_cache-66e96455-fbe8-4a0d-8fe2-9ad60d5abade" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.082363] env[62368]: DEBUG nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Instance network_info: |[{"id": "b78fda7d-cbff-4736-9fcf-49ad36e29f67", "address": "fa:16:3e:c0:9d:3c", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78fda7d-cb", "ovs_interfaceid": "b78fda7d-cbff-4736-9fcf-49ad36e29f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 984.082792] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:9d:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8f40f5c4-c146-449c-884d-6f884dcf2acf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b78fda7d-cbff-4736-9fcf-49ad36e29f67', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.090705] env[62368]: DEBUG oslo.service.loopingcall [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.090940] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 984.091181] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84f1228c-764a-46c7-a1a4-5007c686989e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.113872] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.113872] env[62368]: value = "task-1198937" [ 984.113872] env[62368]: _type = "Task" [ 984.113872] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.122952] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198937, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.148832] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6837ef-3621-418c-bc25-a3f7c8ffd565 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.178382] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.178694] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67acc716-b46b-4c39-b7d3-7f727560d9d9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.187219] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 984.187219] env[62368]: value = "task-1198938" [ 984.187219] env[62368]: _type = "Task" [ 984.187219] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.198943] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198938, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.273839] env[62368]: INFO nova.compute.manager [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Took 16.01 seconds to build instance. [ 984.341579] env[62368]: DEBUG oslo_concurrency.lockutils [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.341712] env[62368]: DEBUG oslo_concurrency.lockutils [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.416238] env[62368]: DEBUG nova.compute.manager [req-6452e0d3-4634-4c78-833b-5fd907eb3be5 req-fd8ce742-27f0-4b39-8de1-2872d510d010 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Received event network-changed-b78fda7d-cbff-4736-9fcf-49ad36e29f67 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.416446] env[62368]: DEBUG nova.compute.manager [req-6452e0d3-4634-4c78-833b-5fd907eb3be5 req-fd8ce742-27f0-4b39-8de1-2872d510d010 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Refreshing instance network info cache due to event network-changed-b78fda7d-cbff-4736-9fcf-49ad36e29f67. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 984.417169] env[62368]: DEBUG oslo_concurrency.lockutils [req-6452e0d3-4634-4c78-833b-5fd907eb3be5 req-fd8ce742-27f0-4b39-8de1-2872d510d010 service nova] Acquiring lock "refresh_cache-66e96455-fbe8-4a0d-8fe2-9ad60d5abade" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.417169] env[62368]: DEBUG oslo_concurrency.lockutils [req-6452e0d3-4634-4c78-833b-5fd907eb3be5 req-fd8ce742-27f0-4b39-8de1-2872d510d010 service nova] Acquired lock "refresh_cache-66e96455-fbe8-4a0d-8fe2-9ad60d5abade" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.417169] env[62368]: DEBUG nova.network.neutron [req-6452e0d3-4634-4c78-833b-5fd907eb3be5 req-fd8ce742-27f0-4b39-8de1-2872d510d010 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Refreshing network info cache for port b78fda7d-cbff-4736-9fcf-49ad36e29f67 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.519028] env[62368]: DEBUG nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.625915] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198937, 'name': CreateVM_Task, 'duration_secs': 0.373425} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.626132] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 984.626793] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.626964] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.627343] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.627602] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f840b170-62cb-4412-932f-9cdd3b6c230c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.632860] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 984.632860] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ff68a1-85a8-90b5-b63b-57121eff8f66" [ 984.632860] env[62368]: _type = "Task" [ 984.632860] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.640855] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ff68a1-85a8-90b5-b63b-57121eff8f66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.696931] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198938, 'name': PowerOffVM_Task, 'duration_secs': 0.253081} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.697230] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.702367] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfiguring VM instance instance-00000058 to detach disk 2002 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 984.703031] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbec785b-3159-4760-a211-8f9a19657193 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.721115] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 984.721115] env[62368]: value = "task-1198939" [ 984.721115] env[62368]: _type = "Task" [ 984.721115] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.729481] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198939, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.777264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a15f144-066b-4a74-b761-01d77a21269b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.522s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.845300] env[62368]: DEBUG nova.compute.utils [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.920602] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.328s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.925507] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.912s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.925739] env[62368]: DEBUG nova.objects.instance [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'resources' on Instance uuid 5633292c-9a74-4c2d-893b-6823f4a587a2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.040046] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.147672] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ff68a1-85a8-90b5-b63b-57121eff8f66, 'name': SearchDatastore_Task, 'duration_secs': 0.010166} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.147996] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.148301] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.148558] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.148713] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.148893] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.149187] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bcdc1caa-5637-4965-91a6-f4bd0706291d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.170924] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.171134] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 985.171908] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55311390-f6f9-4de8-a6f4-bd1db6395b94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.177975] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 985.177975] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527870e3-0f6a-c6ec-b0be-7cd220a64c90" [ 985.177975] env[62368]: _type = "Task" [ 985.177975] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.186704] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527870e3-0f6a-c6ec-b0be-7cd220a64c90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.234881] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198939, 'name': ReconfigVM_Task, 'duration_secs': 0.28986} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.234881] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfigured VM instance instance-00000058 to detach disk 2002 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 985.234881] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 985.234881] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12a44703-546e-49cf-a077-c87605677ca1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.244172] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 985.244172] env[62368]: value = "task-1198940" [ 985.244172] env[62368]: _type = "Task" [ 985.244172] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.256983] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198940, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.350809] env[62368]: DEBUG oslo_concurrency.lockutils [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.501115] env[62368]: INFO nova.scheduler.client.report [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted allocation for migration e3cd6a40-1f21-4511-adbd-66cdf0420c64 [ 985.573688] env[62368]: DEBUG nova.network.neutron [req-6452e0d3-4634-4c78-833b-5fd907eb3be5 req-fd8ce742-27f0-4b39-8de1-2872d510d010 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Updated VIF entry in instance network info cache for port b78fda7d-cbff-4736-9fcf-49ad36e29f67. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 985.574064] env[62368]: DEBUG nova.network.neutron [req-6452e0d3-4634-4c78-833b-5fd907eb3be5 req-fd8ce742-27f0-4b39-8de1-2872d510d010 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Updating instance_info_cache with network_info: [{"id": "b78fda7d-cbff-4736-9fcf-49ad36e29f67", "address": "fa:16:3e:c0:9d:3c", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb78fda7d-cb", "ovs_interfaceid": "b78fda7d-cbff-4736-9fcf-49ad36e29f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.688629] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527870e3-0f6a-c6ec-b0be-7cd220a64c90, 'name': SearchDatastore_Task, 'duration_secs': 0.011457} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.692529] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5170972-1b6b-4ae8-80b5-95714cf338f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.698793] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 985.698793] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526b13cb-a40f-98be-a1af-47e9694777a1" [ 985.698793] env[62368]: _type = "Task" [ 985.698793] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.703134] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4869d54f-a5ac-4997-aa5d-938b719c7257 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.711593] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526b13cb-a40f-98be-a1af-47e9694777a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.714518] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768702eb-b1c5-4eb0-9672-9c52f5570d04 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.752019] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635071a5-bfab-450a-aabd-f7fb1878686f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.763042] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3b518e-b53c-43bf-811e-ba84e2502ae9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.767016] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198940, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.778030] env[62368]: DEBUG nova.compute.provider_tree [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.012782] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2f4ab180-23fc-4e59-92cb-3adc46312d0e tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.444s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.078328] env[62368]: DEBUG oslo_concurrency.lockutils [req-6452e0d3-4634-4c78-833b-5fd907eb3be5 req-fd8ce742-27f0-4b39-8de1-2872d510d010 service nova] Releasing lock "refresh_cache-66e96455-fbe8-4a0d-8fe2-9ad60d5abade" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.113687] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "eea21546-fbbf-4440-829c-8583c4ccabb6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.113975] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.114224] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.114479] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.114679] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.117131] env[62368]: INFO nova.compute.manager [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Terminating instance [ 986.119759] env[62368]: DEBUG nova.compute.manager [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 986.120084] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 986.121318] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f44634-c75c-48fe-9ab2-2537b1b6f9b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.131387] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 986.131654] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91137373-9d4e-4b10-bdde-0eed6e1d0a7c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.138239] env[62368]: DEBUG oslo_vmware.api [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 986.138239] env[62368]: value = "task-1198941" [ 986.138239] env[62368]: _type = "Task" [ 986.138239] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.148209] env[62368]: DEBUG oslo_vmware.api [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.210600] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526b13cb-a40f-98be-a1af-47e9694777a1, 'name': SearchDatastore_Task, 'duration_secs': 0.018647} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.210902] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.211191] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 66e96455-fbe8-4a0d-8fe2-9ad60d5abade/66e96455-fbe8-4a0d-8fe2-9ad60d5abade.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 986.211488] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9da50bc3-23ed-4c69-9d88-923109c7a824 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.219677] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 986.219677] env[62368]: value = "task-1198942" [ 986.219677] env[62368]: _type = "Task" [ 986.219677] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.229265] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198942, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.262297] env[62368]: DEBUG oslo_vmware.api [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1198940, 'name': PowerOnVM_Task, 'duration_secs': 0.718346} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.262567] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 986.262794] env[62368]: DEBUG nova.compute.manager [None req-751bb7b7-56b2-4a0b-8b37-dbb7c38d1241 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.264466] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c6aed3-742d-43c3-afc8-42ba9fdd2142 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.281682] env[62368]: DEBUG nova.scheduler.client.report [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.438878] env[62368]: DEBUG oslo_concurrency.lockutils [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.439479] env[62368]: DEBUG oslo_concurrency.lockutils [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.440082] env[62368]: INFO nova.compute.manager [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Attaching volume c77431e9-8136-407c-9491-6f6caa1d9991 to /dev/sdb [ 986.444698] env[62368]: DEBUG nova.compute.manager [req-b4f4dbf7-2153-444f-a680-0c392dae2adb req-77222a61-7375-438d-9376-225303af38af service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Received event network-changed-76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.445190] env[62368]: DEBUG nova.compute.manager [req-b4f4dbf7-2153-444f-a680-0c392dae2adb req-77222a61-7375-438d-9376-225303af38af service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Refreshing instance network info cache due to event network-changed-76aee038-386b-401a-8c49-5ee204436ab3. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 986.445604] env[62368]: DEBUG oslo_concurrency.lockutils [req-b4f4dbf7-2153-444f-a680-0c392dae2adb req-77222a61-7375-438d-9376-225303af38af service nova] Acquiring lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.445970] env[62368]: DEBUG oslo_concurrency.lockutils [req-b4f4dbf7-2153-444f-a680-0c392dae2adb req-77222a61-7375-438d-9376-225303af38af service nova] Acquired lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.446225] env[62368]: DEBUG nova.network.neutron [req-b4f4dbf7-2153-444f-a680-0c392dae2adb req-77222a61-7375-438d-9376-225303af38af service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Refreshing network info cache for port 76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 986.488277] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c9e690-7cab-4c3d-8e97-a0159f967c15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.496496] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed830f1-58a7-4032-90bb-334428e22390 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.512103] env[62368]: DEBUG nova.virt.block_device [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating existing volume attachment record: 8869db86-ab00-4768-9a55-ef8aef743de6 {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 986.650510] env[62368]: DEBUG oslo_vmware.api [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198941, 'name': PowerOffVM_Task, 'duration_secs': 0.293165} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.650847] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 986.651067] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 986.651753] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de9aa4ca-1f36-42fb-a5a7-e3df9798244f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.731447] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198942, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.733405] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 986.733676] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 986.733897] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleting the datastore file [datastore1] eea21546-fbbf-4440-829c-8583c4ccabb6 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.734305] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d4581c0-bf5a-43cb-8306-970390e3cbca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.741260] env[62368]: DEBUG oslo_vmware.api [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for the task: (returnval){ [ 986.741260] env[62368]: value = "task-1198945" [ 986.741260] env[62368]: _type = "Task" [ 986.741260] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.754694] env[62368]: DEBUG oslo_vmware.api [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198945, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.758232] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.758539] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.758692] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.758879] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.759079] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.761282] env[62368]: INFO nova.compute.manager [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Terminating instance [ 986.763203] env[62368]: DEBUG nova.compute.manager [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 986.763433] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 986.764343] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8e8eb1-0383-4c1f-9a1d-2e6ea07a6005 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.771439] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 986.771827] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd03d52e-25ae-4d01-a813-51e6cce6f48e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.780470] env[62368]: DEBUG oslo_vmware.api [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 986.780470] env[62368]: value = "task-1198946" [ 986.780470] env[62368]: _type = "Task" [ 986.780470] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.787401] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.793466] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.754s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.795273] env[62368]: INFO nova.compute.claims [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.798227] env[62368]: DEBUG oslo_vmware.api [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198946, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.809451] env[62368]: INFO nova.scheduler.client.report [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleted allocations for instance 5633292c-9a74-4c2d-893b-6823f4a587a2 [ 987.231440] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198942, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529108} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.231735] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 66e96455-fbe8-4a0d-8fe2-9ad60d5abade/66e96455-fbe8-4a0d-8fe2-9ad60d5abade.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 987.232012] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.232932] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aea593e4-44ec-4f7a-a077-69ba18a32515 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.239188] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 987.239188] env[62368]: value = "task-1198949" [ 987.239188] env[62368]: _type = "Task" [ 987.239188] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.266140] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.266140] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.266140] env[62368]: DEBUG oslo_vmware.api [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Task: {'id': task-1198945, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.370046} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.266140] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.266140] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.266140] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 987.266749] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 987.266749] env[62368]: INFO nova.compute.manager [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Took 1.15 seconds to destroy the instance on the hypervisor. [ 987.266749] env[62368]: DEBUG oslo.service.loopingcall [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.267248] env[62368]: DEBUG nova.compute.manager [-] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.267248] env[62368]: DEBUG nova.network.neutron [-] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 987.291621] env[62368]: DEBUG oslo_vmware.api [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198946, 'name': PowerOffVM_Task, 'duration_secs': 0.243466} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.291790] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 987.291957] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 987.292234] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8509ecb2-1bc3-44e4-8dc8-b175ddc9d3d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.317481] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d7f5ee96-b3fd-4a4d-b476-c27fa7c70bd6 tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "5633292c-9a74-4c2d-893b-6823f4a587a2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.698s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.535198] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 987.535446] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 987.535634] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleting the datastore file [datastore2] 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.538653] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2f3fc26-102b-4d93-b540-a08b8afd830b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.543264] env[62368]: DEBUG oslo_vmware.api [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 987.543264] env[62368]: value = "task-1198951" [ 987.543264] env[62368]: _type = "Task" [ 987.543264] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.553377] env[62368]: DEBUG oslo_vmware.api [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198951, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.562193] env[62368]: DEBUG nova.network.neutron [req-b4f4dbf7-2153-444f-a680-0c392dae2adb req-77222a61-7375-438d-9376-225303af38af service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Updated VIF entry in instance network info cache for port 76aee038-386b-401a-8c49-5ee204436ab3. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 987.562598] env[62368]: DEBUG nova.network.neutron [req-b4f4dbf7-2153-444f-a680-0c392dae2adb req-77222a61-7375-438d-9376-225303af38af service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Updating instance_info_cache with network_info: [{"id": "76aee038-386b-401a-8c49-5ee204436ab3", "address": "fa:16:3e:46:83:dc", "network": {"id": "0cf66c3e-c52d-4989-98f3-6f97fac0e8a7", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-659832088-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e31b8f1352574bb7808b06e732da7e4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76aee038-38", "ovs_interfaceid": "76aee038-386b-401a-8c49-5ee204436ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.749443] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.412073} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.749710] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.750505] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa5d943-2e95-451b-9acd-3d3c9b3e3f6f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.766254] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "a5cbadbd-20dd-4514-8867-20243af5db0c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.766549] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.766801] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.767044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.767232] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.777524] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 66e96455-fbe8-4a0d-8fe2-9ad60d5abade/66e96455-fbe8-4a0d-8fe2-9ad60d5abade.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.778132] env[62368]: INFO nova.compute.manager [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Terminating instance [ 987.779679] env[62368]: DEBUG nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 987.781982] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2a66f90-98f9-4719-adb9-7f08798b4856 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.797547] env[62368]: DEBUG nova.compute.manager [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 987.797547] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 987.800315] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d2c888-1531-4d62-93c5-535d6b2d8313 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.808739] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 987.808739] env[62368]: value = "task-1198952" [ 987.808739] env[62368]: _type = "Task" [ 987.808739] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.811815] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 987.815521] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d7dffed-c197-4071-bf95-ebda9fd013a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.825036] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198952, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.827339] env[62368]: DEBUG oslo_vmware.api [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 987.827339] env[62368]: value = "task-1198953" [ 987.827339] env[62368]: _type = "Task" [ 987.827339] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.843733] env[62368]: DEBUG oslo_vmware.api [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.054347] env[62368]: DEBUG oslo_vmware.api [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198951, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.379441} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.057420] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.057711] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 988.057933] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 988.058187] env[62368]: INFO nova.compute.manager [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Took 1.29 seconds to destroy the instance on the hypervisor. [ 988.058477] env[62368]: DEBUG oslo.service.loopingcall [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.058929] env[62368]: DEBUG nova.compute.manager [-] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.059046] env[62368]: DEBUG nova.network.neutron [-] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 988.064950] env[62368]: DEBUG oslo_concurrency.lockutils [req-b4f4dbf7-2153-444f-a680-0c392dae2adb req-77222a61-7375-438d-9376-225303af38af service nova] Releasing lock "refresh_cache-ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.070544] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60eb5af6-aafd-4376-acc4-952fb3c4aac7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.078522] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47ec08a-34bc-43f1-aadd-46350fc17e12 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.114309] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f4a9116-9600-46c9-b7c2-31ca897f8fbc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.122949] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3317676-460e-4c6c-8af6-a36de9070f7e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.140747] env[62368]: DEBUG nova.compute.provider_tree [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.217055] env[62368]: DEBUG nova.network.neutron [-] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.323381] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198952, 'name': ReconfigVM_Task, 'duration_secs': 0.382024} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.323901] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 66e96455-fbe8-4a0d-8fe2-9ad60d5abade/66e96455-fbe8-4a0d-8fe2-9ad60d5abade.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.324661] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a30c132-5e1c-4aca-8fb5-17d6f619df43 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.327358] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.335550] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 988.335550] env[62368]: value = "task-1198954" [ 988.335550] env[62368]: _type = "Task" [ 988.335550] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.342448] env[62368]: DEBUG oslo_vmware.api [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198953, 'name': PowerOffVM_Task, 'duration_secs': 0.246686} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.343033] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 988.343250] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 988.343509] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80f121a8-ba62-417e-9824-41bf330bc872 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.348085] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198954, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.413532] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 988.413807] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 988.414012] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleting the datastore file [datastore2] a5cbadbd-20dd-4514-8867-20243af5db0c {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 988.414293] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f78bcbef-7722-4029-87e1-5b0ec2a1ac02 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.421748] env[62368]: DEBUG oslo_vmware.api [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for the task: (returnval){ [ 988.421748] env[62368]: value = "task-1198956" [ 988.421748] env[62368]: _type = "Task" [ 988.421748] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.429970] env[62368]: DEBUG oslo_vmware.api [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198956, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.476740] env[62368]: DEBUG nova.compute.manager [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Received event network-changed-561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.477038] env[62368]: DEBUG nova.compute.manager [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Refreshing instance network info cache due to event network-changed-561cb794-4c95-416d-825a-193df8ddc3ca. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.477316] env[62368]: DEBUG oslo_concurrency.lockutils [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] Acquiring lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.477499] env[62368]: DEBUG oslo_concurrency.lockutils [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] Acquired lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.477698] env[62368]: DEBUG nova.network.neutron [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Refreshing network info cache for port 561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.662426] env[62368]: ERROR nova.scheduler.client.report [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [req-342fdd36-8c4a-4617-a4a4-1642951ba4f5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-342fdd36-8c4a-4617-a4a4-1642951ba4f5"}]} [ 988.683333] env[62368]: DEBUG nova.scheduler.client.report [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 988.698180] env[62368]: DEBUG nova.scheduler.client.report [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 988.698530] env[62368]: DEBUG nova.compute.provider_tree [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.711666] env[62368]: DEBUG nova.scheduler.client.report [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 988.720080] env[62368]: INFO nova.compute.manager [-] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Took 1.45 seconds to deallocate network for instance. [ 988.730427] env[62368]: DEBUG nova.scheduler.client.report [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 988.846917] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198954, 'name': Rename_Task, 'duration_secs': 0.165471} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.847296] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 988.847529] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c24f4ce3-514a-41a0-b734-b7dfb56e47c9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.860100] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 988.860100] env[62368]: value = "task-1198957" [ 988.860100] env[62368]: _type = "Task" [ 988.860100] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.871559] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198957, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.880267] env[62368]: DEBUG nova.network.neutron [-] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.939018] env[62368]: DEBUG oslo_vmware.api [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198956, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.942537] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29001069-e8b3-403f-b476-5fe28a819818 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.952692] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111d0788-e909-49f5-9c09-d0eff6565608 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.987563] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd142010-a905-4bf4-bef0-d642b78fe9e1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.997803] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8de25b-8259-4f87-8f03-01ca5e8afa75 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.010531] env[62368]: DEBUG nova.compute.provider_tree [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.227952] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.240935] env[62368]: DEBUG nova.network.neutron [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updated VIF entry in instance network info cache for port 561cb794-4c95-416d-825a-193df8ddc3ca. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.241410] env[62368]: DEBUG nova.network.neutron [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating instance_info_cache with network_info: [{"id": "561cb794-4c95-416d-825a-193df8ddc3ca", "address": "fa:16:3e:2b:4e:03", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap561cb794-4c", "ovs_interfaceid": "561cb794-4c95-416d-825a-193df8ddc3ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.371873] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198957, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.383086] env[62368]: INFO nova.compute.manager [-] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Took 1.32 seconds to deallocate network for instance. [ 989.433432] env[62368]: DEBUG oslo_vmware.api [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Task: {'id': task-1198956, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.66508} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.433821] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.434019] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 989.434198] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 989.434401] env[62368]: INFO nova.compute.manager [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Took 1.64 seconds to destroy the instance on the hypervisor. [ 989.434637] env[62368]: DEBUG oslo.service.loopingcall [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.434848] env[62368]: DEBUG nova.compute.manager [-] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.434961] env[62368]: DEBUG nova.network.neutron [-] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 989.551409] env[62368]: DEBUG nova.scheduler.client.report [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 129 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 989.551694] env[62368]: DEBUG nova.compute.provider_tree [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 129 to 130 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 989.552098] env[62368]: DEBUG nova.compute.provider_tree [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.744107] env[62368]: DEBUG oslo_concurrency.lockutils [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] Releasing lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.744395] env[62368]: DEBUG nova.compute.manager [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Received event network-changed-561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.744567] env[62368]: DEBUG nova.compute.manager [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Refreshing instance network info cache due to event network-changed-561cb794-4c95-416d-825a-193df8ddc3ca. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.744775] env[62368]: DEBUG oslo_concurrency.lockutils [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] Acquiring lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.744923] env[62368]: DEBUG oslo_concurrency.lockutils [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] Acquired lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.745098] env[62368]: DEBUG nova.network.neutron [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Refreshing network info cache for port 561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.871823] env[62368]: DEBUG oslo_vmware.api [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198957, 'name': PowerOnVM_Task, 'duration_secs': 0.846178} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.872210] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 989.872376] env[62368]: INFO nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Took 8.01 seconds to spawn the instance on the hypervisor. [ 989.872574] env[62368]: DEBUG nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 989.873409] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2e2200-90c5-4c9d-af65-d82aa05f1ea3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.889650] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.050044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.050044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.050241] env[62368]: INFO nova.compute.manager [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Shelving [ 990.056414] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.263s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.056948] env[62368]: DEBUG nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.059888] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.733s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.061272] env[62368]: INFO nova.compute.claims [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.222209] env[62368]: DEBUG nova.network.neutron [-] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.391805] env[62368]: INFO nova.compute.manager [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Took 12.99 seconds to build instance. [ 990.483850] env[62368]: DEBUG nova.network.neutron [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updated VIF entry in instance network info cache for port 561cb794-4c95-416d-825a-193df8ddc3ca. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.484052] env[62368]: DEBUG nova.network.neutron [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating instance_info_cache with network_info: [{"id": "561cb794-4c95-416d-825a-193df8ddc3ca", "address": "fa:16:3e:2b:4e:03", "network": {"id": "0995dd37-fb36-46d0-917b-343fef2d2d4a", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-201275013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3940819fc14a4c628aacd8820efe084e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap561cb794-4c", "ovs_interfaceid": "561cb794-4c95-416d-825a-193df8ddc3ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.504210] env[62368]: DEBUG nova.compute.manager [req-281e3552-2f03-4ab9-bc4b-e89d720a7a53 req-1c9c524c-033e-4561-bde2-c86796cfbcca service nova] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Received event network-vif-deleted-c06f647d-2b3f-4ee0-8221-d00fda253d7b {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.558177] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.558453] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06f170ac-2b41-4a05-b8bd-dd8d1726068e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.568763] env[62368]: DEBUG nova.compute.utils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 990.571037] env[62368]: DEBUG nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 990.571151] env[62368]: DEBUG nova.network.neutron [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 990.573671] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 990.573671] env[62368]: value = "task-1198959" [ 990.573671] env[62368]: _type = "Task" [ 990.573671] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.590181] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198959, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.618976] env[62368]: DEBUG nova.policy [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7dceb8c8f8b4c4dbdde4d097009d429', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b70fcc6664f47d7b55447210851c4cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 990.724845] env[62368]: INFO nova.compute.manager [-] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Took 1.29 seconds to deallocate network for instance. [ 990.894190] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5a1bb4fc-86a4-4273-ad2b-a8eacbb3cdda tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.514s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.957898] env[62368]: DEBUG nova.network.neutron [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Successfully created port: 958ece78-0a1a-4499-ab02-18b1f0732364 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.986588] env[62368]: DEBUG oslo_concurrency.lockutils [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] Releasing lock "refresh_cache-3e09be20-ce79-4bfa-bf84-2877f4e534e6" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.986888] env[62368]: DEBUG nova.compute.manager [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Received event network-vif-deleted-b2ebe893-6ba3-40da-8def-fb6cf1ab442e {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.987160] env[62368]: DEBUG nova.compute.manager [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Received event network-vif-deleted-d229b53b-0e94-41c9-b8e9-0b0eaa366d3c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.987355] env[62368]: INFO nova.compute.manager [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Neutron deleted interface d229b53b-0e94-41c9-b8e9-0b0eaa366d3c; detaching it from the instance and deleting it from the info cache [ 990.987533] env[62368]: DEBUG nova.network.neutron [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.074556] env[62368]: DEBUG nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 991.087668] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198959, 'name': PowerOffVM_Task, 'duration_secs': 0.364643} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.087927] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.088716] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38208c1e-9171-4676-ad3f-df6fad88103a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.117043] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d93ff2-431c-4314-86e6-7d674c658ee4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.232068] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.310702] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c505f50-f20c-4e40-a581-1283ceef9e07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.321021] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba30e63-0f36-4d9d-b043-9051bacf7c1c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.352989] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05f810f-db6f-4f1a-9fec-f7ed8f29805e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.360669] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c0264e-fb83-4c27-b1bf-e522144b5dcd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.375937] env[62368]: DEBUG nova.compute.provider_tree [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.485409] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.485516] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.485794] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.486088] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.486318] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.488835] env[62368]: INFO nova.compute.manager [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Terminating instance [ 991.490848] env[62368]: DEBUG nova.compute.manager [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.491088] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 991.491369] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fb49ac4-ffea-4dd0-a140-532459647f61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.494832] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84886aa5-a3f6-487e-b68f-89d19a4f9d15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.502736] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 991.503899] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8139d844-dba1-44fa-99d7-2585844ef34c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.508340] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ce6e19-6b33-450b-b8be-1444fad591bb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.524878] env[62368]: DEBUG oslo_vmware.api [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 991.524878] env[62368]: value = "task-1198960" [ 991.524878] env[62368]: _type = "Task" [ 991.524878] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.532863] env[62368]: DEBUG oslo_vmware.api [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.542069] env[62368]: DEBUG nova.compute.manager [req-cf83ab6a-9f91-4138-aeb6-ddba147e97d0 req-caf6531f-5202-4723-abf8-42d8eb3f2f4f service nova] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Detach interface failed, port_id=d229b53b-0e94-41c9-b8e9-0b0eaa366d3c, reason: Instance 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 991.568428] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 991.568668] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259881', 'volume_id': 'c77431e9-8136-407c-9491-6f6caa1d9991', 'name': 'volume-c77431e9-8136-407c-9491-6f6caa1d9991', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e6dd9825-f841-49ac-a75a-83d9de3858ac', 'attached_at': '', 'detached_at': '', 'volume_id': 'c77431e9-8136-407c-9491-6f6caa1d9991', 'serial': 'c77431e9-8136-407c-9491-6f6caa1d9991'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 991.569555] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3716649-bfc6-46a4-9117-d9b0f23c45c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.589050] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5137537-6b74-4f78-8252-8b69d97a540f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.616186] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] volume-c77431e9-8136-407c-9491-6f6caa1d9991/volume-c77431e9-8136-407c-9491-6f6caa1d9991.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.616529] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0da5ae2b-cbfb-4e69-a12c-8e937fb210e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.630953] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 991.631502] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1db342b8-fb20-4cba-af94-3b28cc0d3f3a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.640114] env[62368]: DEBUG oslo_vmware.api [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 991.640114] env[62368]: value = "task-1198961" [ 991.640114] env[62368]: _type = "Task" [ 991.640114] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.641437] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 991.641437] env[62368]: value = "task-1198962" [ 991.641437] env[62368]: _type = "Task" [ 991.641437] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.653909] env[62368]: DEBUG oslo_vmware.api [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198961, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.656998] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198962, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.878239] env[62368]: DEBUG nova.scheduler.client.report [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.035734] env[62368]: DEBUG oslo_vmware.api [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.093479] env[62368]: DEBUG nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.121387] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.121679] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.121866] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.122156] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.122269] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.122458] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.122933] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.122933] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.123093] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.123283] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.123495] env[62368]: DEBUG nova.virt.hardware [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.124427] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899f2b92-e734-4ecc-9868-1669145c46c8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.133225] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ddcf64-7a56-4f8d-abb0-b16944edcf33 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.169579] env[62368]: DEBUG oslo_vmware.api [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198961, 'name': ReconfigVM_Task, 'duration_secs': 0.505876} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.173246] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfigured VM instance instance-0000005c to attach disk [datastore2] volume-c77431e9-8136-407c-9491-6f6caa1d9991/volume-c77431e9-8136-407c-9491-6f6caa1d9991.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.179479] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198962, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.179839] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df5ba223-114f-4ba6-b8d4-eb79cdf9d11b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.200125] env[62368]: DEBUG oslo_vmware.api [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 992.200125] env[62368]: value = "task-1198963" [ 992.200125] env[62368]: _type = "Task" [ 992.200125] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.213560] env[62368]: DEBUG oslo_vmware.api [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198963, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.384044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.384944] env[62368]: DEBUG nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 992.388167] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.160s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.388442] env[62368]: DEBUG nova.objects.instance [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lazy-loading 'resources' on Instance uuid eea21546-fbbf-4440-829c-8583c4ccabb6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.526054] env[62368]: DEBUG nova.network.neutron [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Successfully updated port: 958ece78-0a1a-4499-ab02-18b1f0732364 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.532797] env[62368]: DEBUG nova.compute.manager [req-1b118288-0a5f-4023-a426-86bae1155aa5 req-542fa663-a574-43a9-850c-284a64f52e21 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Received event network-vif-plugged-958ece78-0a1a-4499-ab02-18b1f0732364 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.533406] env[62368]: DEBUG oslo_concurrency.lockutils [req-1b118288-0a5f-4023-a426-86bae1155aa5 req-542fa663-a574-43a9-850c-284a64f52e21 service nova] Acquiring lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.533719] env[62368]: DEBUG oslo_concurrency.lockutils [req-1b118288-0a5f-4023-a426-86bae1155aa5 req-542fa663-a574-43a9-850c-284a64f52e21 service nova] Lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.533902] env[62368]: DEBUG oslo_concurrency.lockutils [req-1b118288-0a5f-4023-a426-86bae1155aa5 req-542fa663-a574-43a9-850c-284a64f52e21 service nova] Lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.534094] env[62368]: DEBUG nova.compute.manager [req-1b118288-0a5f-4023-a426-86bae1155aa5 req-542fa663-a574-43a9-850c-284a64f52e21 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] No waiting events found dispatching network-vif-plugged-958ece78-0a1a-4499-ab02-18b1f0732364 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.534271] env[62368]: WARNING nova.compute.manager [req-1b118288-0a5f-4023-a426-86bae1155aa5 req-542fa663-a574-43a9-850c-284a64f52e21 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Received unexpected event network-vif-plugged-958ece78-0a1a-4499-ab02-18b1f0732364 for instance with vm_state building and task_state spawning. [ 992.540993] env[62368]: DEBUG oslo_vmware.api [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198960, 'name': PowerOffVM_Task, 'duration_secs': 0.903242} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.541250] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 992.541419] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 992.541973] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c810ba2a-65e5-43d0-8407-8d16e4b32610 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.606023] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 992.606241] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 992.606433] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleting the datastore file [datastore2] 66e96455-fbe8-4a0d-8fe2-9ad60d5abade {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.606710] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bcbc6783-8db7-4a89-9bb9-bfdc6d239606 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.613481] env[62368]: DEBUG oslo_vmware.api [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 992.613481] env[62368]: value = "task-1198965" [ 992.613481] env[62368]: _type = "Task" [ 992.613481] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.623960] env[62368]: DEBUG oslo_vmware.api [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198965, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.657826] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198962, 'name': CreateSnapshot_Task, 'duration_secs': 0.581978} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.658145] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 992.658868] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03775c2b-8de9-4a87-81cb-57551ffa6775 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.711277] env[62368]: DEBUG oslo_vmware.api [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198963, 'name': ReconfigVM_Task, 'duration_secs': 0.173134} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.711671] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259881', 'volume_id': 'c77431e9-8136-407c-9491-6f6caa1d9991', 'name': 'volume-c77431e9-8136-407c-9491-6f6caa1d9991', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e6dd9825-f841-49ac-a75a-83d9de3858ac', 'attached_at': '', 'detached_at': '', 'volume_id': 'c77431e9-8136-407c-9491-6f6caa1d9991', 'serial': 'c77431e9-8136-407c-9491-6f6caa1d9991'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 992.892085] env[62368]: DEBUG nova.compute.utils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.896780] env[62368]: DEBUG nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 992.897136] env[62368]: DEBUG nova.network.neutron [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 992.942352] env[62368]: DEBUG nova.policy [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0800ab273ca04fbf9396175b57eed6b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e821059910b4e32aab596c6f4d521d8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 993.029775] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-ffa476fb-fb7f-4f68-a38a-85dc9c009857" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.029925] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-ffa476fb-fb7f-4f68-a38a-85dc9c009857" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.030089] env[62368]: DEBUG nova.network.neutron [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 993.114605] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375ac363-2373-4680-bbb1-ee0c7604c0de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.129115] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c32370-5b5e-4d48-9267-8b8b52acdf9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.132489] env[62368]: DEBUG oslo_vmware.api [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1198965, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.467961} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.132857] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 993.133175] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 993.133771] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 993.133771] env[62368]: INFO nova.compute.manager [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Took 1.64 seconds to destroy the instance on the hypervisor. [ 993.133972] env[62368]: DEBUG oslo.service.loopingcall [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.134569] env[62368]: DEBUG nova.compute.manager [-] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 993.134724] env[62368]: DEBUG nova.network.neutron [-] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 993.163916] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20c54d3-7aba-4723-8fcc-6c911f8450f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.178405] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 993.179272] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8df25e49-9265-455e-9f99-6e35f97576b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.185326] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b46e326-234b-4df9-88a8-fedbc8d38f6d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.202639] env[62368]: DEBUG nova.compute.provider_tree [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.211561] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 993.211561] env[62368]: value = "task-1198966" [ 993.211561] env[62368]: _type = "Task" [ 993.211561] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.225024] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198966, 'name': CloneVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.391751] env[62368]: DEBUG nova.network.neutron [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Successfully created port: 35593251-80fd-40d0-8dda-7800a2e2f3ff {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 993.399245] env[62368]: DEBUG nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 993.580787] env[62368]: DEBUG nova.network.neutron [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 993.713548] env[62368]: DEBUG nova.scheduler.client.report [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.734733] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198966, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.737194] env[62368]: DEBUG nova.network.neutron [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Updating instance_info_cache with network_info: [{"id": "958ece78-0a1a-4499-ab02-18b1f0732364", "address": "fa:16:3e:33:05:17", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap958ece78-0a", "ovs_interfaceid": "958ece78-0a1a-4499-ab02-18b1f0732364", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.753205] env[62368]: DEBUG nova.objects.instance [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'flavor' on Instance uuid e6dd9825-f841-49ac-a75a-83d9de3858ac {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.128694] env[62368]: DEBUG nova.network.neutron [-] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.221839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.834s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.224291] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.335s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.224492] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.226724] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.995s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.226975] env[62368]: DEBUG nova.objects.instance [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lazy-loading 'resources' on Instance uuid a5cbadbd-20dd-4514-8867-20243af5db0c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.237022] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198966, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.239494] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-ffa476fb-fb7f-4f68-a38a-85dc9c009857" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.239494] env[62368]: DEBUG nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Instance network_info: |[{"id": "958ece78-0a1a-4499-ab02-18b1f0732364", "address": "fa:16:3e:33:05:17", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap958ece78-0a", "ovs_interfaceid": "958ece78-0a1a-4499-ab02-18b1f0732364", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 994.239878] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:05:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '958ece78-0a1a-4499-ab02-18b1f0732364', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.247460] env[62368]: DEBUG oslo.service.loopingcall [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.248659] env[62368]: INFO nova.scheduler.client.report [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted allocations for instance 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2 [ 994.250151] env[62368]: INFO nova.scheduler.client.report [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Deleted allocations for instance eea21546-fbbf-4440-829c-8583c4ccabb6 [ 994.251152] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 994.253450] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e147536-a112-45f3-ba8a-e70d8498f91b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.273329] env[62368]: DEBUG oslo_concurrency.lockutils [None req-41b25085-39b1-4cf5-a519-07a352879a79 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.834s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.278674] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.278674] env[62368]: value = "task-1198967" [ 994.278674] env[62368]: _type = "Task" [ 994.278674] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.291034] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198967, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.407925] env[62368]: DEBUG nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 994.434565] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.434928] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.435404] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.435679] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.435960] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.436197] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.436653] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.436690] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.436865] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.437039] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.437229] env[62368]: DEBUG nova.virt.hardware [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.439235] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3264ad14-a7e9-4b42-b415-76027ab62a33 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.448354] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e7c7ba-0109-40e8-88c7-b696e2c8213b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.562163] env[62368]: DEBUG nova.compute.manager [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Received event network-changed-958ece78-0a1a-4499-ab02-18b1f0732364 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.562367] env[62368]: DEBUG nova.compute.manager [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Refreshing instance network info cache due to event network-changed-958ece78-0a1a-4499-ab02-18b1f0732364. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 994.562548] env[62368]: DEBUG oslo_concurrency.lockutils [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] Acquiring lock "refresh_cache-ffa476fb-fb7f-4f68-a38a-85dc9c009857" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.562845] env[62368]: DEBUG oslo_concurrency.lockutils [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] Acquired lock "refresh_cache-ffa476fb-fb7f-4f68-a38a-85dc9c009857" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.562932] env[62368]: DEBUG nova.network.neutron [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Refreshing network info cache for port 958ece78-0a1a-4499-ab02-18b1f0732364 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 994.631836] env[62368]: INFO nova.compute.manager [-] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Took 1.50 seconds to deallocate network for instance. [ 994.736589] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198966, 'name': CloneVM_Task, 'duration_secs': 1.449725} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.737123] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Created linked-clone VM from snapshot [ 994.737879] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3ae5ab-b63a-44cd-b501-7d2bf0d9c2ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.746558] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Uploading image 3b257738-fa51-43dd-83ad-cb3309fa62c0 {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 994.768498] env[62368]: DEBUG oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 994.768498] env[62368]: value = "vm-259883" [ 994.768498] env[62368]: _type = "VirtualMachine" [ 994.768498] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 994.768785] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-96febe6a-df17-49c7-8c53-e811ecac4433 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.777200] env[62368]: DEBUG oslo_concurrency.lockutils [None req-07e92def-fb9b-4f2f-99c7-e818536a6780 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.018s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.778393] env[62368]: DEBUG oslo_concurrency.lockutils [None req-f0ce402f-5b4f-4214-8f28-2c846dd8379c tempest-ServersNegativeTestJSON-1844697847 tempest-ServersNegativeTestJSON-1844697847-project-member] Lock "eea21546-fbbf-4440-829c-8583c4ccabb6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.664s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.783764] env[62368]: DEBUG oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lease: (returnval){ [ 994.783764] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5250c2ba-b30e-f586-fca9-e2ed3e3fe894" [ 994.783764] env[62368]: _type = "HttpNfcLease" [ 994.783764] env[62368]: } obtained for exporting VM: (result){ [ 994.783764] env[62368]: value = "vm-259883" [ 994.783764] env[62368]: _type = "VirtualMachine" [ 994.783764] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 994.783972] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the lease: (returnval){ [ 994.783972] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5250c2ba-b30e-f586-fca9-e2ed3e3fe894" [ 994.783972] env[62368]: _type = "HttpNfcLease" [ 994.783972] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 994.793362] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198967, 'name': CreateVM_Task, 'duration_secs': 0.34834} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.794633] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 994.795351] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.795526] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.795864] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 994.797790] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96d7b547-894c-46da-a313-3280862781a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.799317] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 994.799317] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5250c2ba-b30e-f586-fca9-e2ed3e3fe894" [ 994.799317] env[62368]: _type = "HttpNfcLease" [ 994.799317] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 994.802343] env[62368]: DEBUG oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 994.802343] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5250c2ba-b30e-f586-fca9-e2ed3e3fe894" [ 994.802343] env[62368]: _type = "HttpNfcLease" [ 994.802343] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 994.803723] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e6c9c9-cd15-44df-80b9-9e48339266e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.807883] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 994.807883] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c1b4bb-69ac-ea77-3622-ca54f4ff90ea" [ 994.807883] env[62368]: _type = "Task" [ 994.807883] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.814261] env[62368]: DEBUG oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288f1d3-5312-d2b8-9a63-4af06a728ca7/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 994.814442] env[62368]: DEBUG oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288f1d3-5312-d2b8-9a63-4af06a728ca7/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 994.880612] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c1b4bb-69ac-ea77-3622-ca54f4ff90ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.910837] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-102e4eb4-b17e-458f-9cf6-35e85ca126cd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.979401] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f84bbe6-bf08-4fe4-8c48-70898557af14 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.986936] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16596a0a-2175-4cc5-b86c-971d00a466f4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.018494] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a203de14-9f9a-488e-aff6-573695f3a7aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.026689] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e010c426-8894-49de-be15-a8c495d7c2ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.039743] env[62368]: DEBUG nova.compute.provider_tree [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.141659] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.203898] env[62368]: DEBUG nova.compute.manager [req-6ca4c6c4-f379-4fed-adbc-98fe65de9b21 req-57e9395d-5e20-4d39-9399-f3ed6295837d service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Received event network-vif-plugged-35593251-80fd-40d0-8dda-7800a2e2f3ff {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.204207] env[62368]: DEBUG oslo_concurrency.lockutils [req-6ca4c6c4-f379-4fed-adbc-98fe65de9b21 req-57e9395d-5e20-4d39-9399-f3ed6295837d service nova] Acquiring lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.204540] env[62368]: DEBUG oslo_concurrency.lockutils [req-6ca4c6c4-f379-4fed-adbc-98fe65de9b21 req-57e9395d-5e20-4d39-9399-f3ed6295837d service nova] Lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.204785] env[62368]: DEBUG oslo_concurrency.lockutils [req-6ca4c6c4-f379-4fed-adbc-98fe65de9b21 req-57e9395d-5e20-4d39-9399-f3ed6295837d service nova] Lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.205056] env[62368]: DEBUG nova.compute.manager [req-6ca4c6c4-f379-4fed-adbc-98fe65de9b21 req-57e9395d-5e20-4d39-9399-f3ed6295837d service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] No waiting events found dispatching network-vif-plugged-35593251-80fd-40d0-8dda-7800a2e2f3ff {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.205300] env[62368]: WARNING nova.compute.manager [req-6ca4c6c4-f379-4fed-adbc-98fe65de9b21 req-57e9395d-5e20-4d39-9399-f3ed6295837d service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Received unexpected event network-vif-plugged-35593251-80fd-40d0-8dda-7800a2e2f3ff for instance with vm_state building and task_state spawning. [ 995.303389] env[62368]: DEBUG nova.network.neutron [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Updated VIF entry in instance network info cache for port 958ece78-0a1a-4499-ab02-18b1f0732364. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 995.303771] env[62368]: DEBUG nova.network.neutron [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Updating instance_info_cache with network_info: [{"id": "958ece78-0a1a-4499-ab02-18b1f0732364", "address": "fa:16:3e:33:05:17", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap958ece78-0a", "ovs_interfaceid": "958ece78-0a1a-4499-ab02-18b1f0732364", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.320257] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c1b4bb-69ac-ea77-3622-ca54f4ff90ea, 'name': SearchDatastore_Task, 'duration_secs': 0.023271} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.320865] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.321038] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.321338] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.321556] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.321858] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.322308] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d221aa2-0760-4381-a0a3-c7ac45dca81a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.331153] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.331341] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.332088] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e2d5c67-b640-4fbb-beb7-72137f7ebdde {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.340302] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 995.340302] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52b86b0e-3a64-c189-7398-7f0d7f4b5a83" [ 995.340302] env[62368]: _type = "Task" [ 995.340302] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.349039] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b86b0e-3a64-c189-7398-7f0d7f4b5a83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.545208] env[62368]: DEBUG nova.scheduler.client.report [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.639690] env[62368]: DEBUG nova.compute.manager [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Stashing vm_state: active {{(pid=62368) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 995.808377] env[62368]: DEBUG oslo_concurrency.lockutils [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] Releasing lock "refresh_cache-ffa476fb-fb7f-4f68-a38a-85dc9c009857" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.808478] env[62368]: DEBUG nova.compute.manager [req-afbe6de3-296c-4ef7-a0fe-96cc9ce76d22 req-2a3d4874-8533-495c-bead-7484d7eb9150 service nova] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Received event network-vif-deleted-b78fda7d-cbff-4736-9fcf-49ad36e29f67 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.847197] env[62368]: DEBUG nova.network.neutron [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Successfully updated port: 35593251-80fd-40d0-8dda-7800a2e2f3ff {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 995.860819] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b86b0e-3a64-c189-7398-7f0d7f4b5a83, 'name': SearchDatastore_Task, 'duration_secs': 0.008265} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.864137] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-575b69a8-b5c8-4227-894b-7cbb3f65189d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.871529] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 995.871529] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522c9738-1e49-016a-c35e-5638c681b446" [ 995.871529] env[62368]: _type = "Task" [ 995.871529] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.880337] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522c9738-1e49-016a-c35e-5638c681b446, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.050166] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.823s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.052752] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.911s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.053040] env[62368]: DEBUG nova.objects.instance [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lazy-loading 'resources' on Instance uuid 66e96455-fbe8-4a0d-8fe2-9ad60d5abade {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.079546] env[62368]: INFO nova.scheduler.client.report [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Deleted allocations for instance a5cbadbd-20dd-4514-8867-20243af5db0c [ 996.165686] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.351433] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "refresh_cache-9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.351433] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "refresh_cache-9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.351433] env[62368]: DEBUG nova.network.neutron [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.387750] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522c9738-1e49-016a-c35e-5638c681b446, 'name': SearchDatastore_Task, 'duration_secs': 0.009831} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.387750] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.387750] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] ffa476fb-fb7f-4f68-a38a-85dc9c009857/ffa476fb-fb7f-4f68-a38a-85dc9c009857.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 996.387750] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-207f8818-d90b-43b4-9324-67b9c47574cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.393592] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 996.393592] env[62368]: value = "task-1198969" [ 996.393592] env[62368]: _type = "Task" [ 996.393592] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.404434] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198969, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.590525] env[62368]: DEBUG oslo_concurrency.lockutils [None req-24337ed3-8406-47d2-8837-358b82e8927c tempest-AttachInterfacesTestJSON-181815545 tempest-AttachInterfacesTestJSON-181815545-project-member] Lock "a5cbadbd-20dd-4514-8867-20243af5db0c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.823s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.745573] env[62368]: DEBUG nova.compute.manager [req-53a95d94-d026-4b13-b108-5f6c379dbe29 req-ad7e25c9-c395-441f-8775-d2a312f97c25 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Received event network-changed-35593251-80fd-40d0-8dda-7800a2e2f3ff {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.748443] env[62368]: DEBUG nova.compute.manager [req-53a95d94-d026-4b13-b108-5f6c379dbe29 req-ad7e25c9-c395-441f-8775-d2a312f97c25 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Refreshing instance network info cache due to event network-changed-35593251-80fd-40d0-8dda-7800a2e2f3ff. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 996.748443] env[62368]: DEBUG oslo_concurrency.lockutils [req-53a95d94-d026-4b13-b108-5f6c379dbe29 req-ad7e25c9-c395-441f-8775-d2a312f97c25 service nova] Acquiring lock "refresh_cache-9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.811842] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622256d3-5100-4a4e-9d55-2d1a205c572e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.823039] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ff7aed-4c59-4da5-9748-388c472b043d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.884239] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea2983a-4da9-4aae-a506-eea9d1332318 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.893034] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1325e094-62b8-4a3e-ba4f-2a5d7e0c6e35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.914650] env[62368]: DEBUG nova.compute.provider_tree [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.921216] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198969, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.922839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "1d5a93f0-1068-4514-b01c-011e95f48498" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.923065] env[62368]: DEBUG oslo_concurrency.lockutils [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.948094] env[62368]: DEBUG nova.network.neutron [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 997.125447] env[62368]: DEBUG nova.network.neutron [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Updating instance_info_cache with network_info: [{"id": "35593251-80fd-40d0-8dda-7800a2e2f3ff", "address": "fa:16:3e:39:3b:67", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35593251-80", "ovs_interfaceid": "35593251-80fd-40d0-8dda-7800a2e2f3ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.412429] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198969, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561805} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.412707] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] ffa476fb-fb7f-4f68-a38a-85dc9c009857/ffa476fb-fb7f-4f68-a38a-85dc9c009857.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.412931] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.413215] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-223fcacc-f835-45ae-962d-20939a52cac1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.417280] env[62368]: DEBUG nova.scheduler.client.report [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.426109] env[62368]: DEBUG nova.compute.utils [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.427391] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 997.427391] env[62368]: value = "task-1198970" [ 997.427391] env[62368]: _type = "Task" [ 997.427391] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.435750] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.627876] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "refresh_cache-9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.628303] env[62368]: DEBUG nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Instance network_info: |[{"id": "35593251-80fd-40d0-8dda-7800a2e2f3ff", "address": "fa:16:3e:39:3b:67", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35593251-80", "ovs_interfaceid": "35593251-80fd-40d0-8dda-7800a2e2f3ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 997.628683] env[62368]: DEBUG oslo_concurrency.lockutils [req-53a95d94-d026-4b13-b108-5f6c379dbe29 req-ad7e25c9-c395-441f-8775-d2a312f97c25 service nova] Acquired lock "refresh_cache-9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.628955] env[62368]: DEBUG nova.network.neutron [req-53a95d94-d026-4b13-b108-5f6c379dbe29 req-ad7e25c9-c395-441f-8775-d2a312f97c25 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Refreshing network info cache for port 35593251-80fd-40d0-8dda-7800a2e2f3ff {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 997.630355] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:3b:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '35593251-80fd-40d0-8dda-7800a2e2f3ff', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.639655] env[62368]: DEBUG oslo.service.loopingcall [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.643148] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 997.643789] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f82b42b9-8f60-4755-8259-61c1b8a10ead {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.669057] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.669057] env[62368]: value = "task-1198971" [ 997.669057] env[62368]: _type = "Task" [ 997.669057] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.678580] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198971, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.922801] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.870s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.926394] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.761s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.928895] env[62368]: DEBUG oslo_concurrency.lockutils [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.943290] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067918} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.944768] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.946160] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e202af-0041-4067-b2b6-229ea9d9209a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.975637] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] ffa476fb-fb7f-4f68-a38a-85dc9c009857/ffa476fb-fb7f-4f68-a38a-85dc9c009857.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.976943] env[62368]: INFO nova.scheduler.client.report [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted allocations for instance 66e96455-fbe8-4a0d-8fe2-9ad60d5abade [ 997.978295] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfe7b01a-ec1d-4065-ba2a-ed56ce866dc2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.002079] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 998.002079] env[62368]: value = "task-1198972" [ 998.002079] env[62368]: _type = "Task" [ 998.002079] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.010320] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198972, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.085962] env[62368]: DEBUG nova.network.neutron [req-53a95d94-d026-4b13-b108-5f6c379dbe29 req-ad7e25c9-c395-441f-8775-d2a312f97c25 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Updated VIF entry in instance network info cache for port 35593251-80fd-40d0-8dda-7800a2e2f3ff. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 998.086558] env[62368]: DEBUG nova.network.neutron [req-53a95d94-d026-4b13-b108-5f6c379dbe29 req-ad7e25c9-c395-441f-8775-d2a312f97c25 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Updating instance_info_cache with network_info: [{"id": "35593251-80fd-40d0-8dda-7800a2e2f3ff", "address": "fa:16:3e:39:3b:67", "network": {"id": "3d539172-b748-4a4e-aec8-fc9a1dae0406", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1816888884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e821059910b4e32aab596c6f4d521d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35593251-80", "ovs_interfaceid": "35593251-80fd-40d0-8dda-7800a2e2f3ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.179305] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1198971, 'name': CreateVM_Task, 'duration_secs': 0.476842} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.179444] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 998.180071] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.180370] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.180814] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 998.181091] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d248f74-09c2-46f9-a43d-c567bc2a7dd0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.185839] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 998.185839] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52db88b8-7313-0af6-fcd9-f9ecb8f0ae20" [ 998.185839] env[62368]: _type = "Task" [ 998.185839] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.194266] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52db88b8-7313-0af6-fcd9-f9ecb8f0ae20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.438672] env[62368]: INFO nova.compute.claims [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 998.499500] env[62368]: DEBUG oslo_concurrency.lockutils [None req-67ea8240-bdd3-485f-be6e-8f1a5a4a0453 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "66e96455-fbe8-4a0d-8fe2-9ad60d5abade" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.014s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.513920] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198972, 'name': ReconfigVM_Task, 'duration_secs': 0.352491} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.514245] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Reconfigured VM instance instance-00000065 to attach disk [datastore2] ffa476fb-fb7f-4f68-a38a-85dc9c009857/ffa476fb-fb7f-4f68-a38a-85dc9c009857.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.515542] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eaa82279-0888-4267-9946-a3bca3f335ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.522018] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 998.522018] env[62368]: value = "task-1198973" [ 998.522018] env[62368]: _type = "Task" [ 998.522018] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.532147] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198973, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.590202] env[62368]: DEBUG oslo_concurrency.lockutils [req-53a95d94-d026-4b13-b108-5f6c379dbe29 req-ad7e25c9-c395-441f-8775-d2a312f97c25 service nova] Releasing lock "refresh_cache-9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.700547] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52db88b8-7313-0af6-fcd9-f9ecb8f0ae20, 'name': SearchDatastore_Task, 'duration_secs': 0.010382} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.700906] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.701139] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 998.701465] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.701650] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.701836] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 998.702129] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fbf0d45d-12d4-471b-a548-34621f231979 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.711101] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.712070] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 998.712070] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbad8246-7aa5-4c45-b45f-e3b7d8b5ba9b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.718048] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 998.718048] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]522b6eb6-e569-6012-1c96-2f261623b6d4" [ 998.718048] env[62368]: _type = "Task" [ 998.718048] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.726629] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522b6eb6-e569-6012-1c96-2f261623b6d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.944814] env[62368]: INFO nova.compute.resource_tracker [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating resource usage from migration 818c9b11-7728-44b9-a2c9-30d7e48036ce [ 998.995788] env[62368]: DEBUG oslo_concurrency.lockutils [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "1d5a93f0-1068-4514-b01c-011e95f48498" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.996105] env[62368]: DEBUG oslo_concurrency.lockutils [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.996348] env[62368]: INFO nova.compute.manager [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Attaching volume 0a2f79ea-9e92-4df1-8903-20d50069dfee to /dev/sdb [ 999.034573] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198973, 'name': Rename_Task, 'duration_secs': 0.238617} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.034881] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 999.035761] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69f79505-6fe1-4436-8823-2065b2e9dd02 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.039520] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94eae696-d918-4fa7-8ff6-5aaca3c5a981 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.047234] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd94ff3-02bc-4797-aaf8-defa7d8ca89b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.050690] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 999.050690] env[62368]: value = "task-1198974" [ 999.050690] env[62368]: _type = "Task" [ 999.050690] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.061728] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198974, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.068941] env[62368]: DEBUG nova.virt.block_device [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updating existing volume attachment record: 666e9593-382f-4f7c-8c2b-fa83a0cb98bb {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 999.140857] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdd084e-244a-4e5c-aa9f-cf0d8fe7e4e1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.150224] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a755d0-93b7-40c7-86a3-fb6953d5cc30 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.197468] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8f05c7-0ee2-411f-b56d-4971bc4c7b5c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.206592] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d29157a-d3d6-4e7c-acb6-fe335718bde6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.224688] env[62368]: DEBUG nova.compute.provider_tree [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.238815] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]522b6eb6-e569-6012-1c96-2f261623b6d4, 'name': SearchDatastore_Task, 'duration_secs': 0.010059} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.238815] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-062dde13-cc2c-4f82-ad37-52b430120a27 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.245614] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 999.245614] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f3538d-e5a6-7e9f-4735-c59799d6c8c1" [ 999.245614] env[62368]: _type = "Task" [ 999.245614] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.255709] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f3538d-e5a6-7e9f-4735-c59799d6c8c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.563236] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198974, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.731657] env[62368]: DEBUG nova.scheduler.client.report [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.757030] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f3538d-e5a6-7e9f-4735-c59799d6c8c1, 'name': SearchDatastore_Task, 'duration_secs': 0.012575} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.757363] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.757662] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1/9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 999.757959] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc41a319-896e-4de4-b927-ee7cf794aeee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.765817] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 999.765817] env[62368]: value = "task-1198978" [ 999.765817] env[62368]: _type = "Task" [ 999.765817] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.775022] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198978, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.064034] env[62368]: DEBUG oslo_vmware.api [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198974, 'name': PowerOnVM_Task, 'duration_secs': 0.566499} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.064232] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1000.064461] env[62368]: INFO nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Took 7.97 seconds to spawn the instance on the hypervisor. [ 1000.064650] env[62368]: DEBUG nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.065501] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6f0ef3-3cbb-4883-9401-3fd5a851bcd2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.189678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "f409466d-273b-4880-9e45-844909723e4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.190052] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "f409466d-273b-4880-9e45-844909723e4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.237118] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.311s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.237375] env[62368]: INFO nova.compute.manager [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Migrating [ 1000.277496] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198978, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504411} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.277765] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1/9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1000.277994] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1000.278286] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6950db1-92eb-4fc3-99de-c1ee8e14b8b7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.283848] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 1000.283848] env[62368]: value = "task-1198979" [ 1000.283848] env[62368]: _type = "Task" [ 1000.283848] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.291954] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198979, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.586078] env[62368]: INFO nova.compute.manager [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Took 15.56 seconds to build instance. [ 1000.694238] env[62368]: DEBUG nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.758052] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.758052] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.758362] env[62368]: DEBUG nova.network.neutron [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.793586] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198979, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089918} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.793586] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.794304] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db9f3c1-686a-40f2-8a0a-fd9b1c4dcdbb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.816509] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1/9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.816797] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c123ae9-82b1-4553-85c6-7eeb3d091770 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.836396] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 1000.836396] env[62368]: value = "task-1198980" [ 1000.836396] env[62368]: _type = "Task" [ 1000.836396] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.844149] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.088550] env[62368]: DEBUG oslo_concurrency.lockutils [None req-37f5b8e7-44d1-4bdd-bae9-3f21f809cbfa tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.072s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.217740] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.217973] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.219611] env[62368]: INFO nova.compute.claims [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.346429] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198980, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.384067] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ef8b86-e88a-4c62-8f4a-4c2d41428b32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.391102] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ca597743-e79c-4b25-bead-16a86b78d3ff tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Suspending the VM {{(pid=62368) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1001.393576] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-641c07a0-ad4c-44a4-9323-b98380a93433 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.399624] env[62368]: DEBUG oslo_vmware.api [None req-ca597743-e79c-4b25-bead-16a86b78d3ff tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1001.399624] env[62368]: value = "task-1198982" [ 1001.399624] env[62368]: _type = "Task" [ 1001.399624] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.409293] env[62368]: DEBUG oslo_vmware.api [None req-ca597743-e79c-4b25-bead-16a86b78d3ff tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198982, 'name': SuspendVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.494153] env[62368]: DEBUG nova.network.neutron [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [{"id": "05119e53-57b5-4b7d-a056-3bb898352179", "address": "fa:16:3e:3b:b9:de", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05119e53-57", "ovs_interfaceid": "05119e53-57b5-4b7d-a056-3bb898352179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.848338] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198980, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.910884] env[62368]: DEBUG oslo_vmware.api [None req-ca597743-e79c-4b25-bead-16a86b78d3ff tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198982, 'name': SuspendVM_Task} progress is 62%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.997316] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.350222] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198980, 'name': ReconfigVM_Task, 'duration_secs': 1.027061} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.350570] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1/9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.351280] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b5da82a-9df8-44ab-8104-03b32b53febe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.357706] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 1002.357706] env[62368]: value = "task-1198983" [ 1002.357706] env[62368]: _type = "Task" [ 1002.357706] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.368855] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198983, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.409792] env[62368]: DEBUG oslo_vmware.api [None req-ca597743-e79c-4b25-bead-16a86b78d3ff tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198982, 'name': SuspendVM_Task, 'duration_secs': 0.673806} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.412771] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ca597743-e79c-4b25-bead-16a86b78d3ff tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Suspended the VM {{(pid=62368) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1002.412963] env[62368]: DEBUG nova.compute.manager [None req-ca597743-e79c-4b25-bead-16a86b78d3ff tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.413997] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c69bae-f2f4-4455-ae6c-a08a5bb6ffdb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.423697] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f05198-d91e-42ce-b3bb-6409e39568f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.430701] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c07faa-3858-4651-a2e9-a8a45aa48802 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.460628] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1cc9e1-2c8d-4194-b29e-819e8211855d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.468808] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bea4736-6dce-4e91-9f6d-1a471ca035ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.484419] env[62368]: DEBUG nova.compute.provider_tree [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.868779] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198983, 'name': Rename_Task, 'duration_secs': 0.189139} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.868779] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1002.869206] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d298179e-932b-4aaa-9ef7-3eba9c6f646b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.876279] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 1002.876279] env[62368]: value = "task-1198984" [ 1002.876279] env[62368]: _type = "Task" [ 1002.876279] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.884709] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198984, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.987958] env[62368]: DEBUG nova.scheduler.client.report [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.119392] env[62368]: DEBUG oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288f1d3-5312-d2b8-9a63-4af06a728ca7/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1003.120300] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537e46bd-584e-4fa5-ace2-e1d91213fbf5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.126738] env[62368]: DEBUG oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288f1d3-5312-d2b8-9a63-4af06a728ca7/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1003.126916] env[62368]: ERROR oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288f1d3-5312-d2b8-9a63-4af06a728ca7/disk-0.vmdk due to incomplete transfer. [ 1003.127269] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b31c513d-f581-4ac4-b60c-6855307a4b8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.134032] env[62368]: DEBUG oslo_vmware.rw_handles [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288f1d3-5312-d2b8-9a63-4af06a728ca7/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1003.134032] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Uploaded image 3b257738-fa51-43dd-83ad-cb3309fa62c0 to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1003.136185] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1003.136430] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-19eefd00-a8ea-4329-9789-eb67b47983f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.142172] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1003.142172] env[62368]: value = "task-1198985" [ 1003.142172] env[62368]: _type = "Task" [ 1003.142172] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.150221] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198985, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.386262] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198984, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.492964] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.493556] env[62368]: DEBUG nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.510193] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f208b7-e0c8-4990-80b5-239caca9d3b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.532164] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance 'e6dd9825-f841-49ac-a75a-83d9de3858ac' progress to 0 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1003.621726] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1003.621988] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1003.622878] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533ab460-3d14-4d5e-90f6-8ade11b45758 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.639669] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1022cf-95b9-465e-8b26-cb54b20de34b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.665312] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-0a2f79ea-9e92-4df1-8903-20d50069dfee/volume-0a2f79ea-9e92-4df1-8903-20d50069dfee.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.665852] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e8dfb80-1ac2-4105-ab9f-486c38ddcec3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.680913] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198985, 'name': Destroy_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.685773] env[62368]: DEBUG oslo_vmware.api [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1003.685773] env[62368]: value = "task-1198986" [ 1003.685773] env[62368]: _type = "Task" [ 1003.685773] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.692728] env[62368]: DEBUG oslo_vmware.api [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198986, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.813249] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.813592] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.813816] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.814029] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.814211] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.817472] env[62368]: INFO nova.compute.manager [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Terminating instance [ 1003.819611] env[62368]: DEBUG nova.compute.manager [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.819849] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1003.820700] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbb444d-49a6-4824-a274-57c69cdb49cc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.828870] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1003.829144] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ccd08cc-2bd4-4d84-867c-446902ae8a2f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.889106] env[62368]: DEBUG oslo_vmware.api [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198984, 'name': PowerOnVM_Task, 'duration_secs': 0.521137} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.889420] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1003.889725] env[62368]: INFO nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Took 9.48 seconds to spawn the instance on the hypervisor. [ 1003.889887] env[62368]: DEBUG nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.890668] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee77541-24d0-412f-9cd5-fdce8a6a291b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.894126] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1003.894326] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1003.894508] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleting the datastore file [datastore2] ffa476fb-fb7f-4f68-a38a-85dc9c009857 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.895076] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93a8d9d4-af3e-4562-952a-c3003c68a0c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.902047] env[62368]: DEBUG oslo_vmware.api [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1003.902047] env[62368]: value = "task-1198988" [ 1003.902047] env[62368]: _type = "Task" [ 1003.902047] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.909743] env[62368]: DEBUG oslo_vmware.api [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198988, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.998388] env[62368]: DEBUG nova.compute.utils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.999890] env[62368]: DEBUG nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1004.000092] env[62368]: DEBUG nova.network.neutron [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1004.038355] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.039097] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1816e41f-5ca4-4998-aada-70de8ff8ff5f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.045671] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1004.045671] env[62368]: value = "task-1198989" [ 1004.045671] env[62368]: _type = "Task" [ 1004.045671] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.052218] env[62368]: DEBUG nova.policy [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c66c425264424e3eaa37c8058e62cb30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '026a3176abc34001ab3506445226777f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.057197] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198989, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.152900] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198985, 'name': Destroy_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.195517] env[62368]: DEBUG oslo_vmware.api [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198986, 'name': ReconfigVM_Task, 'duration_secs': 0.361913} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.195804] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-0a2f79ea-9e92-4df1-8903-20d50069dfee/volume-0a2f79ea-9e92-4df1-8903-20d50069dfee.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.200684] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-725a74fe-f7e9-41e5-8298-a4f73a9a0aa4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.216829] env[62368]: DEBUG oslo_vmware.api [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1004.216829] env[62368]: value = "task-1198990" [ 1004.216829] env[62368]: _type = "Task" [ 1004.216829] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.229227] env[62368]: DEBUG oslo_vmware.api [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198990, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.298828] env[62368]: DEBUG nova.network.neutron [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Successfully created port: 357ccf24-73b7-4279-911f-96d729d6e747 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.412576] env[62368]: INFO nova.compute.manager [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Took 16.11 seconds to build instance. [ 1004.417731] env[62368]: DEBUG oslo_vmware.api [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1198988, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198927} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.417972] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.418173] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1004.418362] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1004.418528] env[62368]: INFO nova.compute.manager [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1004.418771] env[62368]: DEBUG oslo.service.loopingcall [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.418960] env[62368]: DEBUG nova.compute.manager [-] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.419065] env[62368]: DEBUG nova.network.neutron [-] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1004.503214] env[62368]: DEBUG nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1004.556639] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198989, 'name': PowerOffVM_Task, 'duration_secs': 0.214684} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.557800] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1004.558009] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance 'e6dd9825-f841-49ac-a75a-83d9de3858ac' progress to 17 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1004.653122] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198985, 'name': Destroy_Task, 'duration_secs': 1.294495} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.653390] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Destroyed the VM [ 1004.653622] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1004.653914] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-231f20be-0a3e-46cc-962b-c36b845751a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.660435] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1004.660435] env[62368]: value = "task-1198991" [ 1004.660435] env[62368]: _type = "Task" [ 1004.660435] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.668740] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198991, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.726938] env[62368]: DEBUG oslo_vmware.api [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198990, 'name': ReconfigVM_Task, 'duration_secs': 0.157082} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.727304] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1004.731288] env[62368]: DEBUG nova.compute.manager [req-13a155f1-d0b5-4fb3-a933-f184e466d132 req-87c42fc8-9382-4c2f-9563-66272593adf9 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Received event network-vif-deleted-958ece78-0a1a-4499-ab02-18b1f0732364 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.731496] env[62368]: INFO nova.compute.manager [req-13a155f1-d0b5-4fb3-a933-f184e466d132 req-87c42fc8-9382-4c2f-9563-66272593adf9 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Neutron deleted interface 958ece78-0a1a-4499-ab02-18b1f0732364; detaching it from the instance and deleting it from the info cache [ 1004.731657] env[62368]: DEBUG nova.network.neutron [req-13a155f1-d0b5-4fb3-a933-f184e466d132 req-87c42fc8-9382-4c2f-9563-66272593adf9 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.914807] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4cd05de-0c4e-49cf-ab01-6e12c857e252 tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.651s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.065624] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.065875] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.066145] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.066366] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.066521] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.066675] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.066891] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.067900] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.068167] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.068413] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.068588] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.073992] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74107ed0-e5dc-49b1-a853-da6cb28627ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.091351] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1005.091351] env[62368]: value = "task-1198992" [ 1005.091351] env[62368]: _type = "Task" [ 1005.091351] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.099523] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198992, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.171780] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198991, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.210437] env[62368]: DEBUG nova.network.neutron [-] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.235108] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63eb2304-57c6-4fb6-9eeb-2d8e246ffcef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.248582] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e3b42c-5a10-4b4b-b6b8-4a9e9aa1aed1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.294591] env[62368]: DEBUG nova.compute.manager [req-13a155f1-d0b5-4fb3-a933-f184e466d132 req-87c42fc8-9382-4c2f-9563-66272593adf9 service nova] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Detach interface failed, port_id=958ece78-0a1a-4499-ab02-18b1f0732364, reason: Instance ffa476fb-fb7f-4f68-a38a-85dc9c009857 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1005.516511] env[62368]: DEBUG nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.543813] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.544177] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.544430] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.544695] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.544912] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.545157] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.545444] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.546321] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.546321] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.546321] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.546321] env[62368]: DEBUG nova.virt.hardware [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.547278] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b999ff44-d91b-49f7-b72a-4ff7cbd05bd3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.556049] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8601d7df-0ded-4c08-b451-6b34001496e0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.601168] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198992, 'name': ReconfigVM_Task, 'duration_secs': 0.232067} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.601491] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance 'e6dd9825-f841-49ac-a75a-83d9de3858ac' progress to 33 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1005.671077] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198991, 'name': RemoveSnapshot_Task, 'duration_secs': 0.790682} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.671431] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1005.671719] env[62368]: DEBUG nova.compute.manager [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.672502] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e244010-bc9b-4eb2-9b04-51cb0512508a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.713898] env[62368]: INFO nova.compute.manager [-] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Took 1.29 seconds to deallocate network for instance. [ 1005.763503] env[62368]: DEBUG nova.compute.manager [req-52aeaa6c-9a40-436f-bf1d-a1b86b3e898a req-22273df7-8ba4-460d-9897-67b5f259f4e9 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] Received event network-vif-plugged-357ccf24-73b7-4279-911f-96d729d6e747 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.763730] env[62368]: DEBUG oslo_concurrency.lockutils [req-52aeaa6c-9a40-436f-bf1d-a1b86b3e898a req-22273df7-8ba4-460d-9897-67b5f259f4e9 service nova] Acquiring lock "f409466d-273b-4880-9e45-844909723e4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.763974] env[62368]: DEBUG oslo_concurrency.lockutils [req-52aeaa6c-9a40-436f-bf1d-a1b86b3e898a req-22273df7-8ba4-460d-9897-67b5f259f4e9 service nova] Lock "f409466d-273b-4880-9e45-844909723e4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.764195] env[62368]: DEBUG oslo_concurrency.lockutils [req-52aeaa6c-9a40-436f-bf1d-a1b86b3e898a req-22273df7-8ba4-460d-9897-67b5f259f4e9 service nova] Lock "f409466d-273b-4880-9e45-844909723e4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.764374] env[62368]: DEBUG nova.compute.manager [req-52aeaa6c-9a40-436f-bf1d-a1b86b3e898a req-22273df7-8ba4-460d-9897-67b5f259f4e9 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] No waiting events found dispatching network-vif-plugged-357ccf24-73b7-4279-911f-96d729d6e747 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1005.764542] env[62368]: WARNING nova.compute.manager [req-52aeaa6c-9a40-436f-bf1d-a1b86b3e898a req-22273df7-8ba4-460d-9897-67b5f259f4e9 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] Received unexpected event network-vif-plugged-357ccf24-73b7-4279-911f-96d729d6e747 for instance with vm_state building and task_state spawning. [ 1005.772339] env[62368]: DEBUG nova.objects.instance [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lazy-loading 'flavor' on Instance uuid 1d5a93f0-1068-4514-b01c-011e95f48498 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.836914] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.837269] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.837489] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.837685] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.837863] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.841984] env[62368]: INFO nova.compute.manager [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Terminating instance [ 1005.844641] env[62368]: DEBUG nova.compute.manager [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1005.844772] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1005.845791] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcbe5f9-6575-4826-8152-68712b763071 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.854420] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1005.854860] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01142ffb-6c47-439a-b547-a628f6798ca3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.861375] env[62368]: DEBUG oslo_vmware.api [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 1005.861375] env[62368]: value = "task-1198993" [ 1005.861375] env[62368]: _type = "Task" [ 1005.861375] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.869220] env[62368]: DEBUG nova.network.neutron [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Successfully updated port: 357ccf24-73b7-4279-911f-96d729d6e747 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1005.873330] env[62368]: DEBUG oslo_vmware.api [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198993, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.108751] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.109106] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.109169] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.109363] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.109518] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.109672] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.109879] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.110056] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.110495] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.110683] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.110865] env[62368]: DEBUG nova.virt.hardware [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.116246] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1006.116585] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd7463aa-ebbd-42f6-a1bb-4088f09c3977 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.135026] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1006.135026] env[62368]: value = "task-1198994" [ 1006.135026] env[62368]: _type = "Task" [ 1006.135026] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.143265] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198994, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.185607] env[62368]: INFO nova.compute.manager [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Shelve offloading [ 1006.187418] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1006.187868] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-351df331-f148-4ec2-aa5a-e04bd66da3aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.195559] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1006.195559] env[62368]: value = "task-1198995" [ 1006.195559] env[62368]: _type = "Task" [ 1006.195559] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.205800] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1198995, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.222442] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.222709] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.222936] env[62368]: DEBUG nova.objects.instance [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'resources' on Instance uuid ffa476fb-fb7f-4f68-a38a-85dc9c009857 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.277754] env[62368]: DEBUG oslo_concurrency.lockutils [None req-60778b5a-4b97-473d-92b3-9d29174d91f4 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.281s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.371170] env[62368]: DEBUG oslo_vmware.api [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198993, 'name': PowerOffVM_Task, 'duration_secs': 0.195307} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.371489] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1006.371676] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1006.371934] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7bbb80c-0e22-408f-a0d2-a503e6326954 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.374205] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "refresh_cache-f409466d-273b-4880-9e45-844909723e4f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.374348] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "refresh_cache-f409466d-273b-4880-9e45-844909723e4f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.374542] env[62368]: DEBUG nova.network.neutron [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1006.436771] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1006.437024] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1006.437281] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleting the datastore file [datastore2] 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1006.437547] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cae82e2-a82e-4535-bc1e-55766f4361af {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.444231] env[62368]: DEBUG oslo_vmware.api [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for the task: (returnval){ [ 1006.444231] env[62368]: value = "task-1198997" [ 1006.444231] env[62368]: _type = "Task" [ 1006.444231] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.451932] env[62368]: DEBUG oslo_vmware.api [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198997, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.644585] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198994, 'name': ReconfigVM_Task, 'duration_secs': 0.19386} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.645281] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1006.646077] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9e347b-80ac-4388-b39d-b32175731244 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.672227] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac/e6dd9825-f841-49ac-a75a-83d9de3858ac.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.672621] env[62368]: INFO nova.compute.manager [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Rebuilding instance [ 1006.674380] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cf74143-04e0-4e69-b323-9ae84f5b0436 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.692357] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1006.692357] env[62368]: value = "task-1198998" [ 1006.692357] env[62368]: _type = "Task" [ 1006.692357] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.702065] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198998, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.707119] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1006.707201] env[62368]: DEBUG nova.compute.manager [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.707969] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70060b7-79bc-4980-b291-2285bcc7451a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.715911] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.717203] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.717203] env[62368]: DEBUG nova.network.neutron [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1006.733669] env[62368]: DEBUG nova.compute.manager [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.734522] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fefa6a3-0454-45e7-9ab7-f41275141882 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.881388] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b722d305-1027-4267-9461-fc5edc3cb0c2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.889524] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ff175e-b80c-4465-a5e7-ebb8d7a7473e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.919194] env[62368]: DEBUG nova.network.neutron [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1006.921561] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf634b3-0838-4d82-8d2b-651c6fd79de3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.928957] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf812ba-1cc0-4f3e-b10a-0b24d0b7ec3e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.943545] env[62368]: DEBUG nova.compute.provider_tree [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.954832] env[62368]: DEBUG oslo_vmware.api [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Task: {'id': task-1198997, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133938} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.955646] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.955841] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1006.956029] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1006.956205] env[62368]: INFO nova.compute.manager [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1006.956440] env[62368]: DEBUG oslo.service.loopingcall [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.956623] env[62368]: DEBUG nova.compute.manager [-] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1006.956714] env[62368]: DEBUG nova.network.neutron [-] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1007.061941] env[62368]: DEBUG nova.network.neutron [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Updating instance_info_cache with network_info: [{"id": "357ccf24-73b7-4279-911f-96d729d6e747", "address": "fa:16:3e:77:03:f3", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap357ccf24-73", "ovs_interfaceid": "357ccf24-73b7-4279-911f-96d729d6e747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.209309] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1198998, 'name': ReconfigVM_Task, 'duration_secs': 0.317239} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.210822] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfigured VM instance instance-0000005c to attach disk [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac/e6dd9825-f841-49ac-a75a-83d9de3858ac.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.211118] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance 'e6dd9825-f841-49ac-a75a-83d9de3858ac' progress to 50 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1007.246715] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1007.248797] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-477cd707-c81d-4ffa-bcf2-62d323536647 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.257037] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1007.257037] env[62368]: value = "task-1198999" [ 1007.257037] env[62368]: _type = "Task" [ 1007.257037] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.266328] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198999, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.273163] env[62368]: DEBUG nova.compute.manager [req-340b140e-d6a5-4c55-bfc8-b3ee4a318255 req-849c8bfd-9275-4ba7-aabe-a375fda87258 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Received event network-vif-deleted-35593251-80fd-40d0-8dda-7800a2e2f3ff {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.273367] env[62368]: INFO nova.compute.manager [req-340b140e-d6a5-4c55-bfc8-b3ee4a318255 req-849c8bfd-9275-4ba7-aabe-a375fda87258 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Neutron deleted interface 35593251-80fd-40d0-8dda-7800a2e2f3ff; detaching it from the instance and deleting it from the info cache [ 1007.273488] env[62368]: DEBUG nova.network.neutron [req-340b140e-d6a5-4c55-bfc8-b3ee4a318255 req-849c8bfd-9275-4ba7-aabe-a375fda87258 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.448030] env[62368]: DEBUG nova.scheduler.client.report [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.453827] env[62368]: DEBUG nova.network.neutron [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [{"id": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "address": "fa:16:3e:a8:8e:88", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15eed45d-e0", "ovs_interfaceid": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.564628] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "refresh_cache-f409466d-273b-4880-9e45-844909723e4f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.565139] env[62368]: DEBUG nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Instance network_info: |[{"id": "357ccf24-73b7-4279-911f-96d729d6e747", "address": "fa:16:3e:77:03:f3", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap357ccf24-73", "ovs_interfaceid": "357ccf24-73b7-4279-911f-96d729d6e747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1007.565636] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:03:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8f40f5c4-c146-449c-884d-6f884dcf2acf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '357ccf24-73b7-4279-911f-96d729d6e747', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.574556] env[62368]: DEBUG oslo.service.loopingcall [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.575228] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f409466d-273b-4880-9e45-844909723e4f] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1007.575228] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d645ae2-3b03-4683-ae4e-041d2c95f343 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.596338] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.596338] env[62368]: value = "task-1199000" [ 1007.596338] env[62368]: _type = "Task" [ 1007.596338] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.604708] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199000, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.689646] env[62368]: DEBUG nova.network.neutron [-] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.719454] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f75c04a-282d-46c1-9705-b61acdd6e41b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.746821] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa16f6ee-e278-4f06-b047-dc5cd77031ff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.768775] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance 'e6dd9825-f841-49ac-a75a-83d9de3858ac' progress to 67 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1007.776064] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f472e54e-918f-4535-92e5-562655c1c986 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.783379] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1198999, 'name': PowerOffVM_Task, 'duration_secs': 0.357457} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.784585] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1007.789379] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42775874-8936-49a3-a6e3-9990bd4d6e2c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.807143] env[62368]: DEBUG nova.compute.manager [req-3fd33ded-8b0b-4b74-b266-1428c1a93d01 req-c8e5647f-92c5-4347-88c0-ff89e7977c02 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] Received event network-changed-357ccf24-73b7-4279-911f-96d729d6e747 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.807143] env[62368]: DEBUG nova.compute.manager [req-3fd33ded-8b0b-4b74-b266-1428c1a93d01 req-c8e5647f-92c5-4347-88c0-ff89e7977c02 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] Refreshing instance network info cache due to event network-changed-357ccf24-73b7-4279-911f-96d729d6e747. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1007.807143] env[62368]: DEBUG oslo_concurrency.lockutils [req-3fd33ded-8b0b-4b74-b266-1428c1a93d01 req-c8e5647f-92c5-4347-88c0-ff89e7977c02 service nova] Acquiring lock "refresh_cache-f409466d-273b-4880-9e45-844909723e4f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.807143] env[62368]: DEBUG oslo_concurrency.lockutils [req-3fd33ded-8b0b-4b74-b266-1428c1a93d01 req-c8e5647f-92c5-4347-88c0-ff89e7977c02 service nova] Acquired lock "refresh_cache-f409466d-273b-4880-9e45-844909723e4f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.807143] env[62368]: DEBUG nova.network.neutron [req-3fd33ded-8b0b-4b74-b266-1428c1a93d01 req-c8e5647f-92c5-4347-88c0-ff89e7977c02 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] Refreshing network info cache for port 357ccf24-73b7-4279-911f-96d729d6e747 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1007.824625] env[62368]: DEBUG nova.compute.manager [req-340b140e-d6a5-4c55-bfc8-b3ee4a318255 req-849c8bfd-9275-4ba7-aabe-a375fda87258 service nova] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Detach interface failed, port_id=35593251-80fd-40d0-8dda-7800a2e2f3ff, reason: Instance 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1007.856916] env[62368]: INFO nova.compute.manager [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Detaching volume 0a2f79ea-9e92-4df1-8903-20d50069dfee [ 1007.890649] env[62368]: INFO nova.virt.block_device [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Attempting to driver detach volume 0a2f79ea-9e92-4df1-8903-20d50069dfee from mountpoint /dev/sdb [ 1007.890906] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1007.891304] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1007.893008] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0daae9-6f53-4288-9803-5d3020115ef4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.928233] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd2fdd8-0687-4cfb-982b-897e7ffc10ff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.939927] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d222f0d-21a6-4e38-963a-5f2a841d536b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.960734] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.962826] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.965631] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a397b0-82d7-4163-8874-8c4ac9d8b0e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.981073] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] The volume has not been displaced from its original location: [datastore2] volume-0a2f79ea-9e92-4df1-8903-20d50069dfee/volume-0a2f79ea-9e92-4df1-8903-20d50069dfee.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1007.986443] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1007.987530] env[62368]: INFO nova.scheduler.client.report [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted allocations for instance ffa476fb-fb7f-4f68-a38a-85dc9c009857 [ 1007.988471] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25ada53d-3608-4e2a-b9b6-a49db590e98d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.009837] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1008.009837] env[62368]: value = "task-1199001" [ 1008.009837] env[62368]: _type = "Task" [ 1008.009837] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.017499] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199001, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.022070] env[62368]: INFO nova.compute.manager [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Rebuilding instance [ 1008.066685] env[62368]: DEBUG nova.compute.manager [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.067865] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c06136-08ba-49dc-ad59-7a3938cac626 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.105721] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199000, 'name': CreateVM_Task, 'duration_secs': 0.471952} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.106604] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f409466d-273b-4880-9e45-844909723e4f] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1008.107344] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.107535] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.107898] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.108477] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c252ff48-1d65-41cc-bb5d-d1a70ae5952e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.113013] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1008.113013] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52aee83e-088f-8de2-f110-9e4eb00f7831" [ 1008.113013] env[62368]: _type = "Task" [ 1008.113013] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.122150] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52aee83e-088f-8de2-f110-9e4eb00f7831, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.192446] env[62368]: INFO nova.compute.manager [-] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Took 1.24 seconds to deallocate network for instance. [ 1008.315290] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1008.316416] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cec51fb-9383-43f8-a86a-d4620d3a60b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.324708] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1008.324981] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d957e127-807d-4155-8116-f93cd986c8c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.389445] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1008.389805] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1008.390123] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleting the datastore file [datastore2] fa637a2e-a113-4d23-8924-728ba75f202c {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.390526] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f70f5dea-d5d0-417e-bbfd-96e15893effc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.396792] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1008.396792] env[62368]: value = "task-1199003" [ 1008.396792] env[62368]: _type = "Task" [ 1008.396792] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.407400] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.509023] env[62368]: DEBUG oslo_concurrency.lockutils [None req-33daebc0-7c62-40d9-9d41-be6c20e514b6 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "ffa476fb-fb7f-4f68-a38a-85dc9c009857" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.695s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.522704] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199001, 'name': ReconfigVM_Task, 'duration_secs': 0.192396} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.524887] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1008.530167] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55679c46-9a85-49e3-9a6a-b62af576859a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.546502] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1008.546502] env[62368]: value = "task-1199004" [ 1008.546502] env[62368]: _type = "Task" [ 1008.546502] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.555439] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199004, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.577606] env[62368]: DEBUG nova.network.neutron [req-3fd33ded-8b0b-4b74-b266-1428c1a93d01 req-c8e5647f-92c5-4347-88c0-ff89e7977c02 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] Updated VIF entry in instance network info cache for port 357ccf24-73b7-4279-911f-96d729d6e747. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1008.578301] env[62368]: DEBUG nova.network.neutron [req-3fd33ded-8b0b-4b74-b266-1428c1a93d01 req-c8e5647f-92c5-4347-88c0-ff89e7977c02 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] Updating instance_info_cache with network_info: [{"id": "357ccf24-73b7-4279-911f-96d729d6e747", "address": "fa:16:3e:77:03:f3", "network": {"id": "7cd71508-85db-49fd-85c1-b5cf6f88c1ac", "bridge": "br-int", "label": "tempest-ServersTestJSON-840468912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "026a3176abc34001ab3506445226777f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8f40f5c4-c146-449c-884d-6f884dcf2acf", "external-id": "nsx-vlan-transportzone-240", "segmentation_id": 240, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap357ccf24-73", "ovs_interfaceid": "357ccf24-73b7-4279-911f-96d729d6e747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.580484] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1008.580987] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e869adac-4341-409c-aa40-ec700353f1b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.587949] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 1008.587949] env[62368]: value = "task-1199005" [ 1008.587949] env[62368]: _type = "Task" [ 1008.587949] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.598129] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.622979] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52aee83e-088f-8de2-f110-9e4eb00f7831, 'name': SearchDatastore_Task, 'duration_secs': 0.011264} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.623330] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.623577] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.623837] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.624115] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.624327] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.624601] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8bc2fdeb-c607-4b7c-afa0-74d6503ed8be {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.633793] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.634025] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1008.635294] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3783c963-e8f5-4701-9c56-a09c8f502b15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.641186] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1008.641186] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529bc81a-d38a-c503-1b68-a372de86c6e7" [ 1008.641186] env[62368]: _type = "Task" [ 1008.641186] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.662166] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529bc81a-d38a-c503-1b68-a372de86c6e7, 'name': SearchDatastore_Task, 'duration_secs': 0.018601} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.663074] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71503597-9fbf-4d28-a4a5-d1e7d71fd9d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.668621] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1008.668621] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52009404-f1fe-861f-bcbe-ae2caf60922c" [ 1008.668621] env[62368]: _type = "Task" [ 1008.668621] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.676728] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52009404-f1fe-861f-bcbe-ae2caf60922c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.699180] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.699480] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.699728] env[62368]: DEBUG nova.objects.instance [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lazy-loading 'resources' on Instance uuid 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.908175] env[62368]: DEBUG oslo_vmware.api [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198541} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.908443] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.908632] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1008.908811] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1008.937699] env[62368]: INFO nova.scheduler.client.report [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted allocations for instance fa637a2e-a113-4d23-8924-728ba75f202c [ 1009.056706] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199004, 'name': ReconfigVM_Task, 'duration_secs': 0.180801} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.057040] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1009.082654] env[62368]: DEBUG oslo_concurrency.lockutils [req-3fd33ded-8b0b-4b74-b266-1428c1a93d01 req-c8e5647f-92c5-4347-88c0-ff89e7977c02 service nova] Releasing lock "refresh_cache-f409466d-273b-4880-9e45-844909723e4f" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.097779] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199005, 'name': PowerOffVM_Task, 'duration_secs': 0.187865} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.098066] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1009.098908] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1009.099493] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73d1df7b-a7c3-4f7f-94ce-340a90635f68 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.109573] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 1009.109573] env[62368]: value = "task-1199006" [ 1009.109573] env[62368]: _type = "Task" [ 1009.109573] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.117193] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.180436] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52009404-f1fe-861f-bcbe-ae2caf60922c, 'name': SearchDatastore_Task, 'duration_secs': 0.009492} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.183018] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.183018] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] f409466d-273b-4880-9e45-844909723e4f/f409466d-273b-4880-9e45-844909723e4f.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1009.183018] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d502ad66-060e-4063-a76a-b26769e4794e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.187891] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1009.187891] env[62368]: value = "task-1199007" [ 1009.187891] env[62368]: _type = "Task" [ 1009.187891] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.197249] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199007, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.300863] env[62368]: DEBUG nova.compute.manager [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received event network-vif-unplugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.301141] env[62368]: DEBUG oslo_concurrency.lockutils [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.301423] env[62368]: DEBUG oslo_concurrency.lockutils [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] Lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.301606] env[62368]: DEBUG oslo_concurrency.lockutils [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] Lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.301813] env[62368]: DEBUG nova.compute.manager [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] No waiting events found dispatching network-vif-unplugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1009.302163] env[62368]: WARNING nova.compute.manager [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received unexpected event network-vif-unplugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 for instance with vm_state shelved_offloaded and task_state None. [ 1009.302401] env[62368]: DEBUG nova.compute.manager [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received event network-changed-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.302596] env[62368]: DEBUG nova.compute.manager [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Refreshing instance network info cache due to event network-changed-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.302823] env[62368]: DEBUG oslo_concurrency.lockutils [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] Acquiring lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.303015] env[62368]: DEBUG oslo_concurrency.lockutils [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] Acquired lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.303235] env[62368]: DEBUG nova.network.neutron [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Refreshing network info cache for port 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1009.341833] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b4c587-cdaa-44bb-8ba9-e9e9583c5eb5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.350101] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece70e12-5890-439e-b986-970e45d650b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.380476] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ca8d91-8224-4c44-90f5-4b3852a42aac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.387761] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e2e21f-ab40-4597-beee-8bb1405b0b61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.401288] env[62368]: DEBUG nova.compute.provider_tree [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.442335] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.443453] env[62368]: DEBUG nova.network.neutron [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Port 05119e53-57b5-4b7d-a056-3bb898352179 binding to destination host cpu-1 is already ACTIVE {{(pid=62368) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1009.623206] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1009.623206] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1009.623206] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259862', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'name': 'volume-0923c8dd-6948-46a8-a7ce-88b165401227', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1b238e7-3636-47db-b86e-3fcfe917f705', 'attached_at': '', 'detached_at': '', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'serial': '0923c8dd-6948-46a8-a7ce-88b165401227'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1009.624099] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cdc131-4d9a-48f8-830e-d34fbac6d69a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.648028] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87c19fb-361b-4822-8846-07c158a243e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.656531] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9be6042-cb90-430d-b8d8-d23f88f7cd03 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.677879] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f39093-2c9e-497a-95f2-d1eb904366e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.695605] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] The volume has not been displaced from its original location: [datastore1] volume-0923c8dd-6948-46a8-a7ce-88b165401227/volume-0923c8dd-6948-46a8-a7ce-88b165401227.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1009.700720] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Reconfiguring VM instance instance-00000061 to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1009.704201] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d33ce011-e8cd-4028-958a-6503f3239850 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.722658] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199007, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525913} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.723993] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] f409466d-273b-4880-9e45-844909723e4f/f409466d-273b-4880-9e45-844909723e4f.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1009.724245] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1009.724571] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 1009.724571] env[62368]: value = "task-1199008" [ 1009.724571] env[62368]: _type = "Task" [ 1009.724571] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.724759] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06faf7d4-52f5-49f7-8bb6-912f67a8603c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.734995] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199008, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.738215] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1009.738215] env[62368]: value = "task-1199009" [ 1009.738215] env[62368]: _type = "Task" [ 1009.738215] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.746210] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199009, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.801939] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.802191] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.904812] env[62368]: DEBUG nova.scheduler.client.report [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.108941] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.109247] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68ecfe11-a3dc-4947-bf5e-f0e4695c61f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.117077] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1010.117077] env[62368]: value = "task-1199010" [ 1010.117077] env[62368]: _type = "Task" [ 1010.117077] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.125872] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.237692] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199008, 'name': ReconfigVM_Task, 'duration_secs': 0.296068} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.238037] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Reconfigured VM instance instance-00000061 to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1010.243283] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29217e52-cf6d-47e1-94c5-86ae5ccf09ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.264106] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199009, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070389} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.265355] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1010.265711] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 1010.265711] env[62368]: value = "task-1199011" [ 1010.265711] env[62368]: _type = "Task" [ 1010.265711] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.266442] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25534ce8-1b83-42aa-b12a-34e6c0fd98d1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.292965] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] f409466d-273b-4880-9e45-844909723e4f/f409466d-273b-4880-9e45-844909723e4f.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.297468] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-083b3809-3127-403d-bf04-7d419266b73c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.313036] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199011, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.316475] env[62368]: DEBUG nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1010.326370] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1010.326370] env[62368]: value = "task-1199012" [ 1010.326370] env[62368]: _type = "Task" [ 1010.326370] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.336472] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199012, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.368013] env[62368]: DEBUG nova.network.neutron [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updated VIF entry in instance network info cache for port 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.368634] env[62368]: DEBUG nova.network.neutron [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [{"id": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "address": "fa:16:3e:a8:8e:88", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap15eed45d-e0", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.412635] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.713s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.415700] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.973s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.416133] env[62368]: DEBUG nova.objects.instance [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'resources' on Instance uuid fa637a2e-a113-4d23-8924-728ba75f202c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.431074] env[62368]: INFO nova.scheduler.client.report [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Deleted allocations for instance 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1 [ 1010.463148] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.463400] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.463608] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.626833] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1010.627114] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1010.627369] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1010.628184] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680e3e5b-4c99-4c7b-b2f8-865261d39803 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.648496] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada6013d-3be8-4192-afc2-89fc0f7659b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.654760] env[62368]: WARNING nova.virt.vmwareapi.driver [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1010.655065] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.655787] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8cfa3e-7b60-402a-ba0b-0adcb10dabef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.661812] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1010.662144] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e18a8d7-d7a7-47db-bef0-8cd34a632093 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.778101] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199011, 'name': ReconfigVM_Task, 'duration_secs': 0.116212} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.778431] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259862', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'name': 'volume-0923c8dd-6948-46a8-a7ce-88b165401227', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c1b238e7-3636-47db-b86e-3fcfe917f705', 'attached_at': '', 'detached_at': '', 'volume_id': '0923c8dd-6948-46a8-a7ce-88b165401227', 'serial': '0923c8dd-6948-46a8-a7ce-88b165401227'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1010.778716] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.779493] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f87935-c1db-44df-af74-fefe83037f78 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.785779] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1010.785973] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6d176b8-1ae3-4035-ad52-4034df82447c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.842276] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199012, 'name': ReconfigVM_Task, 'duration_secs': 0.317163} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.843381] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.843740] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Reconfigured VM instance instance-00000067 to attach disk [datastore1] f409466d-273b-4880-9e45-844909723e4f/f409466d-273b-4880-9e45-844909723e4f.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1010.844363] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-434411ad-5abe-4f5a-a7ee-733dae840703 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.850707] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1010.850980] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1010.851117] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Deleting the datastore file [datastore1] c1b238e7-3636-47db-b86e-3fcfe917f705 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.851381] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e10a5fd-04d1-4a99-8c2f-a447a963a35d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.855035] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1010.855035] env[62368]: value = "task-1199015" [ 1010.855035] env[62368]: _type = "Task" [ 1010.855035] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.859974] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for the task: (returnval){ [ 1010.859974] env[62368]: value = "task-1199016" [ 1010.859974] env[62368]: _type = "Task" [ 1010.859974] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.867047] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199015, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.870545] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199016, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.870929] env[62368]: DEBUG oslo_concurrency.lockutils [req-216c1623-6f1c-4e7a-bba8-9982c5a52549 req-267c3c4f-6be8-4692-983e-d2cdd3d1efc6 service nova] Releasing lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.919181] env[62368]: DEBUG nova.objects.instance [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'numa_topology' on Instance uuid fa637a2e-a113-4d23-8924-728ba75f202c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.932287] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1010.932426] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1010.932634] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleting the datastore file [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.932948] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b765140-b88d-4682-b98e-689a15f03bb0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.940248] env[62368]: DEBUG oslo_concurrency.lockutils [None req-85bad14e-1dff-4b25-bcab-6579316100be tempest-ServerDiskConfigTestJSON-1630132274 tempest-ServerDiskConfigTestJSON-1630132274-project-member] Lock "9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.103s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.943078] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1010.943078] env[62368]: value = "task-1199017" [ 1010.943078] env[62368]: _type = "Task" [ 1010.943078] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.951447] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199017, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.338011] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.367378] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199015, 'name': Rename_Task, 'duration_secs': 0.180107} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.368036] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1011.368284] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bbf69d2-c1bf-48b4-9bbb-ef5b9f4a08ad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.372589] env[62368]: DEBUG oslo_vmware.api [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Task: {'id': task-1199016, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094841} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.373149] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.373361] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1011.373529] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1011.377421] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1011.377421] env[62368]: value = "task-1199018" [ 1011.377421] env[62368]: _type = "Task" [ 1011.377421] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.387350] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199018, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.422517] env[62368]: DEBUG nova.objects.base [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1011.442216] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1011.442765] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33b3d8e4-1866-4b13-a935-6621affc3486 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.456327] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab021f62-053c-46e0-8967-73d7a279aaa0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.471688] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199017, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150375} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.472671] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.472848] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1011.473101] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1011.493074] env[62368]: ERROR nova.compute.manager [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Failed to detach volume 0923c8dd-6948-46a8-a7ce-88b165401227 from /dev/sda: nova.exception.InstanceNotFound: Instance c1b238e7-3636-47db-b86e-3fcfe917f705 could not be found. [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Traceback (most recent call last): [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self.driver.rebuild(**kwargs) [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] raise NotImplementedError() [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] NotImplementedError [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] During handling of the above exception, another exception occurred: [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Traceback (most recent call last): [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self.driver.detach_volume(context, old_connection_info, [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] return self._volumeops.detach_volume(connection_info, instance) [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self._detach_volume_vmdk(connection_info, instance) [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] stable_ref.fetch_moref(session) [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] nova.exception.InstanceNotFound: Instance c1b238e7-3636-47db-b86e-3fcfe917f705 could not be found. [ 1011.493074] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] [ 1011.507564] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.507848] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.508063] env[62368]: DEBUG nova.network.neutron [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.615406] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71804814-24b6-4868-85ea-77e43e3116a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.619497] env[62368]: DEBUG nova.compute.utils [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Build of instance c1b238e7-3636-47db-b86e-3fcfe917f705 aborted: Failed to rebuild volume backed instance. {{(pid=62368) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1011.622311] env[62368]: ERROR nova.compute.manager [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance c1b238e7-3636-47db-b86e-3fcfe917f705 aborted: Failed to rebuild volume backed instance. [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Traceback (most recent call last): [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self.driver.rebuild(**kwargs) [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] raise NotImplementedError() [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] NotImplementedError [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] During handling of the above exception, another exception occurred: [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Traceback (most recent call last): [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self._detach_root_volume(context, instance, root_bdm) [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] with excutils.save_and_reraise_exception(): [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self.force_reraise() [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] raise self.value [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self.driver.detach_volume(context, old_connection_info, [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] return self._volumeops.detach_volume(connection_info, instance) [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self._detach_volume_vmdk(connection_info, instance) [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] stable_ref.fetch_moref(session) [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] nova.exception.InstanceNotFound: Instance c1b238e7-3636-47db-b86e-3fcfe917f705 could not be found. [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] During handling of the above exception, another exception occurred: [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Traceback (most recent call last): [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] yield [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 1011.622311] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self._do_rebuild_instance_with_claim( [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self._do_rebuild_instance( [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self._rebuild_default_impl(**kwargs) [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] self._rebuild_volume_backed_instance( [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] raise exception.BuildAbortException( [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] nova.exception.BuildAbortException: Build of instance c1b238e7-3636-47db-b86e-3fcfe917f705 aborted: Failed to rebuild volume backed instance. [ 1011.623864] env[62368]: ERROR nova.compute.manager [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] [ 1011.628177] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5aaa31b-ec87-4c2d-844d-90dcfa05090c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.660429] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d06f525-92dc-405e-8e5b-c4475c42f5a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.669565] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f886cddc-f554-4294-9c7e-446f7862373c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.685236] env[62368]: DEBUG nova.compute.provider_tree [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.887288] env[62368]: DEBUG oslo_vmware.api [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199018, 'name': PowerOnVM_Task, 'duration_secs': 0.469451} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.887644] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1011.887745] env[62368]: INFO nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Took 6.37 seconds to spawn the instance on the hypervisor. [ 1011.887903] env[62368]: DEBUG nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.888699] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f739573-11a9-4de3-895f-d56de62b5e5e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.981419] env[62368]: INFO nova.virt.block_device [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Booting with volume 0a2f79ea-9e92-4df1-8903-20d50069dfee at /dev/sdb [ 1012.020507] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7cc3cfa0-05b5-4cc9-8d21-4357309ce827 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.029850] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8802b6-4e58-4130-89b9-1210d1ded910 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.060956] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21241ef2-6863-444d-a6df-39088f5b7d72 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.070525] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2622f3c4-f00b-4922-818a-0d9e4ba019d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.107604] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05542ba8-36ea-48f8-87d3-d443bd0dddda {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.114822] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad44d0e-5253-4472-a924-f56c8a7845d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.131243] env[62368]: DEBUG nova.virt.block_device [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updating existing volume attachment record: 99e8abd4-4a2d-4c16-a144-4e436f74525b {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1012.189271] env[62368]: DEBUG nova.scheduler.client.report [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.316904] env[62368]: DEBUG nova.network.neutron [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [{"id": "05119e53-57b5-4b7d-a056-3bb898352179", "address": "fa:16:3e:3b:b9:de", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05119e53-57", "ovs_interfaceid": "05119e53-57b5-4b7d-a056-3bb898352179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.406025] env[62368]: INFO nova.compute.manager [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Took 11.21 seconds to build instance. [ 1012.695593] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.280s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.698203] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.855s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.699996] env[62368]: INFO nova.compute.claims [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.819730] env[62368]: DEBUG oslo_concurrency.lockutils [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.908354] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a935216c-46d6-4dd9-897f-76390db8ed9d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "f409466d-273b-4880-9e45-844909723e4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.718s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.211600] env[62368]: DEBUG oslo_concurrency.lockutils [None req-993ca035-5cac-4b65-9716-0120f21f5b20 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.161s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.213864] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.874s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.213864] env[62368]: INFO nova.compute.manager [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Unshelving [ 1013.331448] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e85dde-1f04-4470-82e1-8c03b15b1b58 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.339283] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d46e4ec-6928-4f18-8881-a0f525b1dd07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.639362] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.811428] env[62368]: DEBUG oslo_concurrency.lockutils [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "f409466d-273b-4880-9e45-844909723e4f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.811428] env[62368]: DEBUG oslo_concurrency.lockutils [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "f409466d-273b-4880-9e45-844909723e4f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.811428] env[62368]: DEBUG nova.compute.manager [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.811428] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513a8b16-909a-496f-9cb1-3d7d66109c2b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.817612] env[62368]: DEBUG nova.compute.manager [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62368) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1013.818223] env[62368]: DEBUG nova.objects.instance [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lazy-loading 'flavor' on Instance uuid f409466d-273b-4880-9e45-844909723e4f {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.901130] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71da58af-d015-45ea-b39f-f68043cb6053 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.912956] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487348fd-2460-4682-9478-6e212e6cb591 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.944458] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701d8542-ade1-478d-8002-3deb165e71fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.952401] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2c6b62-d359-4be3-9ea1-7bce41d619bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.966518] env[62368]: DEBUG nova.compute.provider_tree [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.227027] env[62368]: DEBUG nova.compute.utils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.252160] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Acquiring lock "c1b238e7-3636-47db-b86e-3fcfe917f705" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.252160] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "c1b238e7-3636-47db-b86e-3fcfe917f705" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.252160] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Acquiring lock "c1b238e7-3636-47db-b86e-3fcfe917f705-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.252160] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "c1b238e7-3636-47db-b86e-3fcfe917f705-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.252306] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "c1b238e7-3636-47db-b86e-3fcfe917f705-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.254774] env[62368]: INFO nova.compute.manager [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Terminating instance [ 1014.258024] env[62368]: DEBUG nova.compute.manager [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1014.258024] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5125665-79e7-43df-8eae-dbb2afbf2769 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.267618] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1ca3cf-ac79-4b2f-a35e-746a6d97879b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.285643] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.285932] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.286100] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.286300] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.286455] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.286606] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.286818] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.286981] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.287191] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.287462] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.287649] env[62368]: DEBUG nova.virt.hardware [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.288414] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28511c3-045d-4f28-bf9c-43550dbb3671 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.301481] env[62368]: WARNING nova.virt.vmwareapi.driver [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance c1b238e7-3636-47db-b86e-3fcfe917f705 could not be found. [ 1014.301640] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1014.302346] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b7b3236-65c5-4336-b7da-579b852d6e59 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.307353] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a33fec-3354-4733-98cd-728670d2445f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.314860] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26511738-f864-487d-bb85-9b86766b94a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.334201] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:07:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd47a86a3-3fd6-48d2-8a80-e6b7f201bd76', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.341755] env[62368]: DEBUG oslo.service.loopingcall [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.342314] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1014.342948] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1014.343168] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bcdbf07a-22b1-4819-98b3-e4425beaeb5f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.344712] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-768cd2c9-19d7-459f-b1d8-0e7b782e4d51 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.369452] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c1b238e7-3636-47db-b86e-3fcfe917f705 could not be found. [ 1014.369452] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1014.369452] env[62368]: INFO nova.compute.manager [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Took 0.11 seconds to destroy the instance on the hypervisor. [ 1014.369452] env[62368]: DEBUG oslo.service.loopingcall [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.370090] env[62368]: DEBUG nova.compute.manager [-] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.370090] env[62368]: DEBUG nova.network.neutron [-] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.373305] env[62368]: DEBUG oslo_vmware.api [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1014.373305] env[62368]: value = "task-1199019" [ 1014.373305] env[62368]: _type = "Task" [ 1014.373305] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.374510] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.374510] env[62368]: value = "task-1199020" [ 1014.374510] env[62368]: _type = "Task" [ 1014.374510] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.385420] env[62368]: DEBUG oslo_vmware.api [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.388420] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199020, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.463449] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825c7eea-fced-4d49-a87c-ead5d85d6b5d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.469463] env[62368]: DEBUG nova.scheduler.client.report [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.493214] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.795s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.493810] env[62368]: DEBUG nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.498210] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03dda69-06ff-410a-9a62-dfe56db0c5fb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.502036] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.863s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.512447] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance 'e6dd9825-f841-49ac-a75a-83d9de3858ac' progress to 83 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1014.704052] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8c1264-cf10-4761-8717-8e6b18e24d78 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.712219] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc503ca-11ed-47dc-a0c9-930a70574699 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.743588] env[62368]: INFO nova.virt.block_device [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Booting with volume 8880c5c4-9ad5-4209-884a-84eb03713365 at /dev/sdb [ 1014.746936] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35dfdf4-256a-45ae-9bb6-18d67f04e1a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.756846] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2488d5a8-96e2-4d8a-a30b-a0bc817c96b7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.770927] env[62368]: DEBUG nova.compute.provider_tree [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.781681] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a81080c0-08ea-4b16-97fc-e8de5b9cc4c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.789904] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27baad9-4fd1-4185-bd74-bd765805d248 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.821784] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a6239551-34f5-4072-90ef-04f0cc3d8a41 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.830042] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e4148e-4c06-4d98-9b7c-3930b83f664f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.862249] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65576cbd-4382-4e78-b7b2-1c7c42831dd3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.868824] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c02a266-f597-4b6e-8021-b576b662fbcf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.884661] env[62368]: DEBUG nova.virt.block_device [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating existing volume attachment record: 07dc05ea-dcbe-470c-8dd1-b566875120dd {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1014.890035] env[62368]: DEBUG oslo_vmware.api [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199019, 'name': PowerOffVM_Task, 'duration_secs': 0.207496} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.890302] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1014.890488] env[62368]: DEBUG nova.compute.manager [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.891259] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d380e3-1c25-4cd0-93d0-4b88d7cb1bea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.896598] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199020, 'name': CreateVM_Task, 'duration_secs': 0.338001} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.897232] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1014.897899] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.898115] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.898473] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.900693] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02ddd779-c6c2-4487-8559-6573dc405657 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.907134] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1014.907134] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52408204-519a-4821-de63-8f0ece17b3e4" [ 1014.907134] env[62368]: _type = "Task" [ 1014.907134] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.914290] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52408204-519a-4821-de63-8f0ece17b3e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.978530] env[62368]: DEBUG nova.compute.manager [req-e2574c3b-cac3-423a-afb2-8ee038cecc9d req-279c0bee-8bc5-442a-969c-ba935d494e0f service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Received event network-vif-deleted-d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.978611] env[62368]: INFO nova.compute.manager [req-e2574c3b-cac3-423a-afb2-8ee038cecc9d req-279c0bee-8bc5-442a-969c-ba935d494e0f service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Neutron deleted interface d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c; detaching it from the instance and deleting it from the info cache [ 1014.979755] env[62368]: DEBUG nova.network.neutron [req-e2574c3b-cac3-423a-afb2-8ee038cecc9d req-279c0bee-8bc5-442a-969c-ba935d494e0f service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.999851] env[62368]: DEBUG nova.compute.utils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.000827] env[62368]: DEBUG nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1015.001010] env[62368]: DEBUG nova.network.neutron [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1015.020135] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1015.020549] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c147a986-6c6e-48ce-8020-accdd4ee697c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.028336] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1015.028336] env[62368]: value = "task-1199022" [ 1015.028336] env[62368]: _type = "Task" [ 1015.028336] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.036840] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199022, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.080819] env[62368]: DEBUG nova.policy [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7dceb8c8f8b4c4dbdde4d097009d429', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b70fcc6664f47d7b55447210851c4cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1015.274034] env[62368]: DEBUG nova.scheduler.client.report [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.407897] env[62368]: DEBUG nova.network.neutron [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Successfully created port: 8702b61c-a9f0-44cc-85e2-3024bb421ef2 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.418558] env[62368]: DEBUG oslo_concurrency.lockutils [None req-931e4330-3723-40de-b2f2-f14fa413abf7 tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "f409466d-273b-4880-9e45-844909723e4f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.609s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.431665] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52408204-519a-4821-de63-8f0ece17b3e4, 'name': SearchDatastore_Task, 'duration_secs': 0.012446} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.431665] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.432049] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.432382] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.432614] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.432840] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.436028] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55ff11e6-5cd1-47df-8290-ebfacff213f4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.436028] env[62368]: DEBUG nova.network.neutron [-] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.443391] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.443612] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1015.444987] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acf10d84-a200-48c4-929a-202190732900 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.451742] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1015.451742] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]525a0383-590f-8ecd-64c5-6141746b58ba" [ 1015.451742] env[62368]: _type = "Task" [ 1015.451742] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.461125] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525a0383-590f-8ecd-64c5-6141746b58ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.482863] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-555a9e3f-96f2-4d0b-9c32-434f1bf3f002 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.491332] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c5233b-1dbe-4c5a-89c2-9f9f2bdb0668 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.504659] env[62368]: DEBUG nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.522594] env[62368]: DEBUG nova.compute.manager [req-e2574c3b-cac3-423a-afb2-8ee038cecc9d req-279c0bee-8bc5-442a-969c-ba935d494e0f service nova] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Detach interface failed, port_id=d4014180-3c8e-4b8a-8a62-9e0f4fbe2c9c, reason: Instance c1b238e7-3636-47db-b86e-3fcfe917f705 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1015.539461] env[62368]: DEBUG oslo_vmware.api [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199022, 'name': PowerOnVM_Task, 'duration_secs': 0.413291} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.539762] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.540061] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-175bd57e-b42d-48c5-96fe-fbfe273e9999 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance 'e6dd9825-f841-49ac-a75a-83d9de3858ac' progress to 100 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1015.781952] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.280s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.782235] env[62368]: INFO nova.compute.manager [None req-fd776887-3a9a-477f-a278-b22ebf280c00 tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Successfully reverted task state from rebuilding on failure for instance. [ 1015.938805] env[62368]: INFO nova.compute.manager [-] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Took 1.57 seconds to deallocate network for instance. [ 1015.962864] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525a0383-590f-8ecd-64c5-6141746b58ba, 'name': SearchDatastore_Task, 'duration_secs': 0.012397} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.963709] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1816439e-6dcb-4673-a5f0-5aff09f2d455 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.969590] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1015.969590] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521792db-1ff4-2ea5-8d13-1cd40302a031" [ 1015.969590] env[62368]: _type = "Task" [ 1015.969590] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.979866] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521792db-1ff4-2ea5-8d13-1cd40302a031, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.484500] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521792db-1ff4-2ea5-8d13-1cd40302a031, 'name': SearchDatastore_Task, 'duration_secs': 0.042853} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.485109] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.485193] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498/1d5a93f0-1068-4514-b01c-011e95f48498.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1016.485458] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7262b3cd-bab6-4394-b5c7-b9df824e5e8c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.492903] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1016.492903] env[62368]: value = "task-1199025" [ 1016.492903] env[62368]: _type = "Task" [ 1016.492903] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.500639] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199025, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.503628] env[62368]: INFO nova.compute.manager [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Took 0.56 seconds to detach 1 volumes for instance. [ 1016.507015] env[62368]: DEBUG nova.compute.manager [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Deleting volume: 0923c8dd-6948-46a8-a7ce-88b165401227 {{(pid=62368) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1016.513845] env[62368]: DEBUG nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1016.542454] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.542745] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.543848] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.543848] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.543848] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.543848] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.543848] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.544452] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.544452] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.544452] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.544580] env[62368]: DEBUG nova.virt.hardware [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.545426] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b5e610-25f3-421d-aa61-5ca59fc9a133 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.555908] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342d165d-0e59-48df-ad9e-e07e74fa402d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.647022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "f409466d-273b-4880-9e45-844909723e4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.647022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "f409466d-273b-4880-9e45-844909723e4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.647022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "f409466d-273b-4880-9e45-844909723e4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.647022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "f409466d-273b-4880-9e45-844909723e4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.647022] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "f409466d-273b-4880-9e45-844909723e4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.653019] env[62368]: INFO nova.compute.manager [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Terminating instance [ 1016.653019] env[62368]: DEBUG nova.compute.manager [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1016.653019] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1016.653019] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279cffba-1b1d-4e1a-aabf-880d05bd8558 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.662658] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1016.663752] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6de09b1-808a-4f8c-a69c-b1bde221c548 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.734536] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1016.734536] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1016.734536] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleting the datastore file [datastore1] f409466d-273b-4880-9e45-844909723e4f {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.734536] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81f0c5c2-d17c-4774-b8c1-f3892ac026aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.743023] env[62368]: DEBUG oslo_vmware.api [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1016.743023] env[62368]: value = "task-1199028" [ 1016.743023] env[62368]: _type = "Task" [ 1016.743023] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.750741] env[62368]: DEBUG oslo_vmware.api [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199028, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.006256] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199025, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485636} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.008767] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498/1d5a93f0-1068-4514-b01c-011e95f48498.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1017.008767] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.008767] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8e932b8-385e-497b-ad80-68456a1d9098 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.016739] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1017.016739] env[62368]: value = "task-1199029" [ 1017.016739] env[62368]: _type = "Task" [ 1017.016739] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.028326] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199029, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.037248] env[62368]: DEBUG nova.compute.manager [req-ab4279de-1d0b-42e8-82aa-90a80b5315e6 req-166ee2bb-25ff-4cef-a371-99afb90d1902 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Received event network-vif-plugged-8702b61c-a9f0-44cc-85e2-3024bb421ef2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.037899] env[62368]: DEBUG oslo_concurrency.lockutils [req-ab4279de-1d0b-42e8-82aa-90a80b5315e6 req-166ee2bb-25ff-4cef-a371-99afb90d1902 service nova] Acquiring lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.037899] env[62368]: DEBUG oslo_concurrency.lockutils [req-ab4279de-1d0b-42e8-82aa-90a80b5315e6 req-166ee2bb-25ff-4cef-a371-99afb90d1902 service nova] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.038062] env[62368]: DEBUG oslo_concurrency.lockutils [req-ab4279de-1d0b-42e8-82aa-90a80b5315e6 req-166ee2bb-25ff-4cef-a371-99afb90d1902 service nova] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.038517] env[62368]: DEBUG nova.compute.manager [req-ab4279de-1d0b-42e8-82aa-90a80b5315e6 req-166ee2bb-25ff-4cef-a371-99afb90d1902 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] No waiting events found dispatching network-vif-plugged-8702b61c-a9f0-44cc-85e2-3024bb421ef2 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1017.038755] env[62368]: WARNING nova.compute.manager [req-ab4279de-1d0b-42e8-82aa-90a80b5315e6 req-166ee2bb-25ff-4cef-a371-99afb90d1902 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Received unexpected event network-vif-plugged-8702b61c-a9f0-44cc-85e2-3024bb421ef2 for instance with vm_state building and task_state spawning. [ 1017.106613] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.106884] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.107129] env[62368]: DEBUG nova.objects.instance [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lazy-loading 'resources' on Instance uuid c1b238e7-3636-47db-b86e-3fcfe917f705 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.136437] env[62368]: DEBUG nova.network.neutron [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Successfully updated port: 8702b61c-a9f0-44cc-85e2-3024bb421ef2 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1017.252394] env[62368]: DEBUG oslo_vmware.api [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199028, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.298818} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.252769] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1017.252972] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1017.253175] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1017.253360] env[62368]: INFO nova.compute.manager [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: f409466d-273b-4880-9e45-844909723e4f] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1017.253964] env[62368]: DEBUG oslo.service.loopingcall [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.254203] env[62368]: DEBUG nova.compute.manager [-] [instance: f409466d-273b-4880-9e45-844909723e4f] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1017.254303] env[62368]: DEBUG nova.network.neutron [-] [instance: f409466d-273b-4880-9e45-844909723e4f] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1017.527059] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199029, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066745} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.527388] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.528145] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0944136-5325-46e6-ad76-ddb73d587eef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.550122] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498/1d5a93f0-1068-4514-b01c-011e95f48498.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.550401] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73bd7136-9425-422b-b219-f6998a68223d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.577035] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1017.577035] env[62368]: value = "task-1199031" [ 1017.577035] env[62368]: _type = "Task" [ 1017.577035] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.584755] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199031, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.642377] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.642539] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.642696] env[62368]: DEBUG nova.network.neutron [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1017.815500] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0bf535-7db6-45d7-9167-b7018800dd21 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.823616] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c31410c-b354-4040-87fb-7a417aa2ac3c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.856592] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d35a7e-4513-40ad-a922-22d2717c43bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.864840] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ba452e-f1ed-4314-a050-3af2b88924c1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.879764] env[62368]: DEBUG nova.compute.provider_tree [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.995439] env[62368]: DEBUG nova.network.neutron [-] [instance: f409466d-273b-4880-9e45-844909723e4f] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.089202] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199031, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.196388] env[62368]: DEBUG nova.network.neutron [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1018.382845] env[62368]: DEBUG nova.scheduler.client.report [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.467603] env[62368]: DEBUG nova.network.neutron [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance_info_cache with network_info: [{"id": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "address": "fa:16:3e:5b:ce:6e", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8702b61c-a9", "ovs_interfaceid": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.497793] env[62368]: INFO nova.compute.manager [-] [instance: f409466d-273b-4880-9e45-844909723e4f] Took 1.24 seconds to deallocate network for instance. [ 1018.589112] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199031, 'name': ReconfigVM_Task, 'duration_secs': 0.574397} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.589472] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498/1d5a93f0-1068-4514-b01c-011e95f48498.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.590857] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'encryption_format': None, 'encrypted': False, 'size': 0, 'device_name': '/dev/sda', 'boot_index': 0, 'disk_bus': None, 'encryption_options': None, 'device_type': 'disk', 'encryption_secret_uuid': None, 'image_id': 'd6bd1b2e-f5f9-49c5-bb6f-9536d441b77d'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'guest_format': None, 'attachment_id': '99e8abd4-4a2d-4c16-a144-4e436f74525b', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'}, 'mount_device': '/dev/sdb', 'boot_index': None, 'disk_bus': None, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=62368) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1018.591105] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1018.591312] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1018.592182] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba86c64-5d35-4012-9e8c-02df5a154bdf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.612990] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427afdd6-5d48-4c77-a2bf-d7c2d8d1c66f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.641531] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-0a2f79ea-9e92-4df1-8903-20d50069dfee/volume-0a2f79ea-9e92-4df1-8903-20d50069dfee.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.641890] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8743cb05-fbdf-4dfe-b213-322f68d5069b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.661266] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1018.661266] env[62368]: value = "task-1199032" [ 1018.661266] env[62368]: _type = "Task" [ 1018.661266] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.670445] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199032, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.711396] env[62368]: DEBUG nova.network.neutron [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Port 05119e53-57b5-4b7d-a056-3bb898352179 binding to destination host cpu-1 is already ACTIVE {{(pid=62368) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1018.711701] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.711865] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.712068] env[62368]: DEBUG nova.network.neutron [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.888459] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.970265] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.970578] env[62368]: DEBUG nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Instance network_info: |[{"id": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "address": "fa:16:3e:5b:ce:6e", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8702b61c-a9", "ovs_interfaceid": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1018.971013] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:ce:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8702b61c-a9f0-44cc-85e2-3024bb421ef2', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.979017] env[62368]: DEBUG oslo.service.loopingcall [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.979289] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1018.979500] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fdf0c691-26e7-4408-a1a8-f9e60bb1a144 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.001243] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.001243] env[62368]: value = "task-1199033" [ 1019.001243] env[62368]: _type = "Task" [ 1019.001243] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.005187] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.005436] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.005656] env[62368]: DEBUG nova.objects.instance [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lazy-loading 'resources' on Instance uuid f409466d-273b-4880-9e45-844909723e4f {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.011723] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199033, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.071885] env[62368]: DEBUG nova.compute.manager [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Received event network-changed-8702b61c-a9f0-44cc-85e2-3024bb421ef2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.072040] env[62368]: DEBUG nova.compute.manager [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Refreshing instance network info cache due to event network-changed-8702b61c-a9f0-44cc-85e2-3024bb421ef2. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1019.072351] env[62368]: DEBUG oslo_concurrency.lockutils [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] Acquiring lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.072549] env[62368]: DEBUG oslo_concurrency.lockutils [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] Acquired lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.072728] env[62368]: DEBUG nova.network.neutron [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Refreshing network info cache for port 8702b61c-a9f0-44cc-85e2-3024bb421ef2 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1019.172225] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199032, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.409463] env[62368]: DEBUG oslo_concurrency.lockutils [None req-8f985069-bfd2-4632-abbc-95ce361b322f tempest-ServerActionsV293TestJSON-74412560 tempest-ServerActionsV293TestJSON-74412560-project-member] Lock "c1b238e7-3636-47db-b86e-3fcfe917f705" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.157s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.524471] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199033, 'name': CreateVM_Task, 'duration_secs': 0.368823} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.525619] env[62368]: DEBUG nova.network.neutron [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [{"id": "05119e53-57b5-4b7d-a056-3bb898352179", "address": "fa:16:3e:3b:b9:de", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05119e53-57", "ovs_interfaceid": "05119e53-57b5-4b7d-a056-3bb898352179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.526867] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1019.527603] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.527796] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.528142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1019.528695] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-076939b8-4eed-4369-96c3-af9618e8c10d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.533671] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1019.533671] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5290997d-8e05-1dd2-1751-5ac9ba858e43" [ 1019.533671] env[62368]: _type = "Task" [ 1019.533671] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.545350] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5290997d-8e05-1dd2-1751-5ac9ba858e43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.672570] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199032, 'name': ReconfigVM_Task, 'duration_secs': 0.610522} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.675019] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-0a2f79ea-9e92-4df1-8903-20d50069dfee/volume-0a2f79ea-9e92-4df1-8903-20d50069dfee.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.686598] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22b07b84-a70a-435a-9aa4-23c5107f425d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.705692] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1019.705692] env[62368]: value = "task-1199034" [ 1019.705692] env[62368]: _type = "Task" [ 1019.705692] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.711736] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fa0e3f-6324-467b-8494-123b16baa771 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.716976] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199034, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.721277] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e06a11-2ced-4c84-a620-9d5675ae1ddf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.752888] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50d4f28-06e0-4fee-9586-d71eae66527e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.759940] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6e183e-715d-4186-aef3-f8c7ae2afb07 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.772982] env[62368]: DEBUG nova.compute.provider_tree [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.825518] env[62368]: DEBUG nova.network.neutron [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updated VIF entry in instance network info cache for port 8702b61c-a9f0-44cc-85e2-3024bb421ef2. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.825972] env[62368]: DEBUG nova.network.neutron [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance_info_cache with network_info: [{"id": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "address": "fa:16:3e:5b:ce:6e", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8702b61c-a9", "ovs_interfaceid": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.029772] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.044819] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5290997d-8e05-1dd2-1751-5ac9ba858e43, 'name': SearchDatastore_Task, 'duration_secs': 0.039132} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.045117] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.045620] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.045620] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.045788] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.045955] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.046261] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57bd487f-c9c2-4281-b54d-443df187d4bc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.058779] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.058953] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1020.059648] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f616870e-fad9-4952-850f-a44c34f55c59 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.064302] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1020.064302] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521f5637-5132-52c6-cae4-0e869cf66d59" [ 1020.064302] env[62368]: _type = "Task" [ 1020.064302] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.072717] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521f5637-5132-52c6-cae4-0e869cf66d59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.215304] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199034, 'name': ReconfigVM_Task, 'duration_secs': 0.141332} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.215648] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1020.216165] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d0ab4ba-2886-489f-a91f-132c2be6355b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.222430] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1020.222430] env[62368]: value = "task-1199035" [ 1020.222430] env[62368]: _type = "Task" [ 1020.222430] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.229949] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199035, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.276269] env[62368]: DEBUG nova.scheduler.client.report [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.328338] env[62368]: DEBUG oslo_concurrency.lockutils [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] Releasing lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.328711] env[62368]: DEBUG nova.compute.manager [req-f3737da9-6963-4238-b0c9-d2a310cdfbf3 req-e8ed2aec-307e-4f8e-86c3-5c91baf7f2d1 service nova] [instance: f409466d-273b-4880-9e45-844909723e4f] Received event network-vif-deleted-357ccf24-73b7-4279-911f-96d729d6e747 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.492145] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.532819] env[62368]: DEBUG nova.compute.manager [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62368) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1020.575802] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521f5637-5132-52c6-cae4-0e869cf66d59, 'name': SearchDatastore_Task, 'duration_secs': 0.055707} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.576602] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bee01ed-c061-49bc-bc26-8d84f3347975 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.581429] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1020.581429] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e64aeb-4475-0195-bf4a-d236e844fe58" [ 1020.581429] env[62368]: _type = "Task" [ 1020.581429] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.591166] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e64aeb-4475-0195-bf4a-d236e844fe58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.732244] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199035, 'name': Rename_Task, 'duration_secs': 0.155576} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.732525] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1020.732811] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1ee3993-a4f5-4542-a25c-c691ae9a294c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.738998] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1020.738998] env[62368]: value = "task-1199036" [ 1020.738998] env[62368]: _type = "Task" [ 1020.738998] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.746051] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.781018] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.783347] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.291s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.783596] env[62368]: DEBUG nova.objects.instance [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'pci_requests' on Instance uuid fa637a2e-a113-4d23-8924-728ba75f202c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.798559] env[62368]: INFO nova.scheduler.client.report [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted allocations for instance f409466d-273b-4880-9e45-844909723e4f [ 1021.092116] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e64aeb-4475-0195-bf4a-d236e844fe58, 'name': SearchDatastore_Task, 'duration_secs': 0.008899} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.092403] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.092665] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] c3dde3bb-380f-40e8-bc39-ffdba6ba3756/c3dde3bb-380f-40e8-bc39-ffdba6ba3756.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1021.092924] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-547e5674-890a-433c-aa9e-fe004166b4e6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.099419] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1021.099419] env[62368]: value = "task-1199037" [ 1021.099419] env[62368]: _type = "Task" [ 1021.099419] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.106746] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.249441] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199036, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.287344] env[62368]: DEBUG nova.objects.instance [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'numa_topology' on Instance uuid fa637a2e-a113-4d23-8924-728ba75f202c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.305320] env[62368]: DEBUG oslo_concurrency.lockutils [None req-91de3b40-1228-4487-b76e-8990fd0c62cd tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "f409466d-273b-4880-9e45-844909723e4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.660s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.610689] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199037, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447115} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.610971] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] c3dde3bb-380f-40e8-bc39-ffdba6ba3756/c3dde3bb-380f-40e8-bc39-ffdba6ba3756.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1021.611211] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.611563] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5fa3c2de-3ff2-47fa-a004-c7e91eb3c19a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.617012] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1021.617012] env[62368]: value = "task-1199038" [ 1021.617012] env[62368]: _type = "Task" [ 1021.617012] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.624948] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199038, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.631800] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.748922] env[62368]: DEBUG oslo_vmware.api [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199036, 'name': PowerOnVM_Task, 'duration_secs': 0.844973} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.749200] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1021.749467] env[62368]: DEBUG nova.compute.manager [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.750305] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba717f3-6d94-410c-b7f5-af28a2ab4fb0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.790714] env[62368]: INFO nova.compute.claims [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.127866] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199038, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065639} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.127866] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.128634] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29467573-965e-408e-ba56-8fb66a51e920 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.150251] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] c3dde3bb-380f-40e8-bc39-ffdba6ba3756/c3dde3bb-380f-40e8-bc39-ffdba6ba3756.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.150514] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fffeb2a-74ea-4204-bb51-1a71ed26759a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.169969] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1022.169969] env[62368]: value = "task-1199039" [ 1022.169969] env[62368]: _type = "Task" [ 1022.169969] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.177856] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199039, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.269445] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.459393] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.459717] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.460024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.460279] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.460512] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.463895] env[62368]: INFO nova.compute.manager [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Terminating instance [ 1022.466506] env[62368]: DEBUG nova.compute.manager [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.466791] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1022.467926] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2890e15-e01a-41a4-984f-2722fb4b0aad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.477558] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.477839] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7beffc29-3919-47d1-a1d4-37d41f69c375 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.485019] env[62368]: DEBUG oslo_vmware.api [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1022.485019] env[62368]: value = "task-1199040" [ 1022.485019] env[62368]: _type = "Task" [ 1022.485019] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.493829] env[62368]: DEBUG oslo_vmware.api [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199040, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.680835] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199039, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.822391] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.822391] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.943572] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f624107b-34f4-4078-bfef-e33c27632aa1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.951933] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8315abdc-4f8f-4743-8f64-09b9e06c414b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.982128] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2b98b2-f403-47d0-9d76-25d80b90b6c8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.991493] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2203203b-90a3-4878-bf1f-705c2c0fe6a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.998304] env[62368]: DEBUG oslo_vmware.api [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199040, 'name': PowerOffVM_Task, 'duration_secs': 0.200325} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.998914] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1022.999113] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1022.999348] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10dbfc1b-a3ec-438e-b0ee-13799ae45d50 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.008113] env[62368]: DEBUG nova.compute.provider_tree [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.074869] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1023.075127] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1023.075440] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleting the datastore file [datastore1] b79f0e79-9e3a-47c7-9949-8743601ec6c5 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.075765] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4294b370-1584-4bd1-88a5-66e1c291b269 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.085432] env[62368]: DEBUG oslo_vmware.api [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for the task: (returnval){ [ 1023.085432] env[62368]: value = "task-1199042" [ 1023.085432] env[62368]: _type = "Task" [ 1023.085432] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.092969] env[62368]: DEBUG oslo_vmware.api [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.181217] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199039, 'name': ReconfigVM_Task, 'duration_secs': 0.623091} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.181545] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Reconfigured VM instance instance-00000068 to attach disk [datastore2] c3dde3bb-380f-40e8-bc39-ffdba6ba3756/c3dde3bb-380f-40e8-bc39-ffdba6ba3756.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.182242] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f1f97a5-22fa-4307-8482-88690fbf31b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.188951] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1023.188951] env[62368]: value = "task-1199043" [ 1023.188951] env[62368]: _type = "Task" [ 1023.188951] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.197699] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199043, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.326862] env[62368]: DEBUG nova.compute.utils [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1023.351862] env[62368]: DEBUG oslo_concurrency.lockutils [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "1d5a93f0-1068-4514-b01c-011e95f48498" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.352127] env[62368]: DEBUG oslo_concurrency.lockutils [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.511413] env[62368]: DEBUG nova.scheduler.client.report [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.595970] env[62368]: DEBUG oslo_vmware.api [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Task: {'id': task-1199042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139952} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.596294] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.596602] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1023.596857] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1023.597053] env[62368]: INFO nova.compute.manager [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1023.597336] env[62368]: DEBUG oslo.service.loopingcall [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.597585] env[62368]: DEBUG nova.compute.manager [-] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.597623] env[62368]: DEBUG nova.network.neutron [-] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1023.699449] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199043, 'name': Rename_Task, 'duration_secs': 0.136224} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.699763] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1023.700060] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f44b452b-ddbf-489c-87d9-98c0b95830f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.707438] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1023.707438] env[62368]: value = "task-1199044" [ 1023.707438] env[62368]: _type = "Task" [ 1023.707438] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.715466] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199044, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.829526] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.855542] env[62368]: INFO nova.compute.manager [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Detaching volume 0a2f79ea-9e92-4df1-8903-20d50069dfee [ 1023.893699] env[62368]: INFO nova.virt.block_device [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Attempting to driver detach volume 0a2f79ea-9e92-4df1-8903-20d50069dfee from mountpoint /dev/sdb [ 1023.893974] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1023.894247] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1023.895205] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f0a498-2356-475d-8a7c-f0c4d4cffbd7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.918118] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e805fa-84e5-40b6-b0e7-22ecbb35650a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.925228] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863e7f27-8d62-4185-9496-1097589a3457 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.947739] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ae32da-8719-4465-acaf-a6179312a897 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.963358] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] The volume has not been displaced from its original location: [datastore2] volume-0a2f79ea-9e92-4df1-8903-20d50069dfee/volume-0a2f79ea-9e92-4df1-8903-20d50069dfee.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1023.968860] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1023.969106] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11acec29-60cc-4dfc-95d0-e6d3c4173d1d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.984057] env[62368]: DEBUG nova.compute.manager [req-806b5e41-430e-440e-9845-78fb2015c062 req-7e07a858-3a3e-4dd3-b5d8-bfbf4667c46e service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Received event network-vif-deleted-bf67cd0c-ff81-451a-8e0e-6ee50bb77f78 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.984266] env[62368]: INFO nova.compute.manager [req-806b5e41-430e-440e-9845-78fb2015c062 req-7e07a858-3a3e-4dd3-b5d8-bfbf4667c46e service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Neutron deleted interface bf67cd0c-ff81-451a-8e0e-6ee50bb77f78; detaching it from the instance and deleting it from the info cache [ 1023.984446] env[62368]: DEBUG nova.network.neutron [req-806b5e41-430e-440e-9845-78fb2015c062 req-7e07a858-3a3e-4dd3-b5d8-bfbf4667c46e service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.991125] env[62368]: DEBUG oslo_vmware.api [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1023.991125] env[62368]: value = "task-1199045" [ 1023.991125] env[62368]: _type = "Task" [ 1023.991125] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.000190] env[62368]: DEBUG oslo_vmware.api [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199045, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.008272] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.008523] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.016287] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.233s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.018555] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 2.387s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.048998] env[62368]: INFO nova.network.neutron [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating port 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1024.220617] env[62368]: DEBUG oslo_vmware.api [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199044, 'name': PowerOnVM_Task, 'duration_secs': 0.443517} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.220617] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1024.220617] env[62368]: INFO nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Took 7.71 seconds to spawn the instance on the hypervisor. [ 1024.220617] env[62368]: DEBUG nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.220617] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9baefd31-66c1-4fef-ade4-5c07db3f9a96 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.367057] env[62368]: DEBUG nova.network.neutron [-] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.487815] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3f78c89f-85df-421b-9059-e43389201b31 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.498086] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c09002-e779-40f0-b777-fd5f9ba0ee2b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.512482] env[62368]: INFO nova.compute.manager [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Detaching volume cf7750c2-86cd-43ff-89ac-880138be3dc5 [ 1024.514115] env[62368]: DEBUG oslo_vmware.api [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199045, 'name': ReconfigVM_Task, 'duration_secs': 0.314214} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.514839] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1024.521171] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e873440a-b3e4-4e64-b348-e643b725c6a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.531095] env[62368]: DEBUG nova.objects.instance [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'migration_context' on Instance uuid e6dd9825-f841-49ac-a75a-83d9de3858ac {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.540958] env[62368]: DEBUG nova.compute.manager [req-806b5e41-430e-440e-9845-78fb2015c062 req-7e07a858-3a3e-4dd3-b5d8-bfbf4667c46e service nova] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Detach interface failed, port_id=bf67cd0c-ff81-451a-8e0e-6ee50bb77f78, reason: Instance b79f0e79-9e3a-47c7-9949-8743601ec6c5 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1024.546968] env[62368]: DEBUG oslo_vmware.api [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1024.546968] env[62368]: value = "task-1199046" [ 1024.546968] env[62368]: _type = "Task" [ 1024.546968] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.554745] env[62368]: DEBUG oslo_vmware.api [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199046, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.555675] env[62368]: INFO nova.virt.block_device [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Attempting to driver detach volume cf7750c2-86cd-43ff-89ac-880138be3dc5 from mountpoint /dev/sdb [ 1024.555883] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1024.556081] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259876', 'volume_id': 'cf7750c2-86cd-43ff-89ac-880138be3dc5', 'name': 'volume-cf7750c2-86cd-43ff-89ac-880138be3dc5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e09be20-ce79-4bfa-bf84-2877f4e534e6', 'attached_at': '', 'detached_at': '', 'volume_id': 'cf7750c2-86cd-43ff-89ac-880138be3dc5', 'serial': 'cf7750c2-86cd-43ff-89ac-880138be3dc5'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1024.556926] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd9034a-d668-419d-925c-9b0d3405ac6c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.577279] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3658c891-be75-40fe-b8fb-44a66ec35cff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.583733] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa30e84-f017-44a6-9d00-a08f47ab1b22 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.603011] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485d8f1b-969b-4d5d-a196-7628796dfbb5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.617258] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] The volume has not been displaced from its original location: [datastore1] volume-cf7750c2-86cd-43ff-89ac-880138be3dc5/volume-cf7750c2-86cd-43ff-89ac-880138be3dc5.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1024.622376] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1024.622643] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d386f6c8-d810-4e7a-a076-3fa8689c29dd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.639053] env[62368]: DEBUG oslo_vmware.api [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 1024.639053] env[62368]: value = "task-1199047" [ 1024.639053] env[62368]: _type = "Task" [ 1024.639053] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.647942] env[62368]: DEBUG oslo_vmware.api [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199047, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.738986] env[62368]: INFO nova.compute.manager [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Took 13.92 seconds to build instance. [ 1024.870054] env[62368]: INFO nova.compute.manager [-] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Took 1.27 seconds to deallocate network for instance. [ 1024.905179] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.905652] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.906783] env[62368]: INFO nova.compute.manager [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Attaching volume 4e212d2b-6fb0-4bbf-9c37-a44e83960938 to /dev/sdb [ 1024.941492] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6060eb5-c462-45e0-b80a-001f6f217a0e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.949435] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438a2e9f-8940-4a08-8a53-d935b10d3aa0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.962870] env[62368]: DEBUG nova.virt.block_device [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Updating existing volume attachment record: cc8fa165-b9e5-44f9-9060-449acc7d22a1 {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1025.056847] env[62368]: DEBUG oslo_vmware.api [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199046, 'name': ReconfigVM_Task, 'duration_secs': 0.201927} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.057427] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259887', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'name': 'volume-0a2f79ea-9e92-4df1-8903-20d50069dfee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1d5a93f0-1068-4514-b01c-011e95f48498', 'attached_at': '', 'detached_at': '', 'volume_id': '0a2f79ea-9e92-4df1-8903-20d50069dfee', 'serial': '0a2f79ea-9e92-4df1-8903-20d50069dfee'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1025.149791] env[62368]: DEBUG oslo_vmware.api [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199047, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.194983] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764fb461-63e0-4d19-9111-c8229d6b7c91 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.204200] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5e6865-30e5-4b26-a21e-bed368fd1b96 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.242508] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9c40669e-cf9f-4f7f-a90f-52c7284bd987 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.440s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.243478] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7451903-e3f1-4b1a-8c27-b79c94a160d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.252762] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66bd5d2-c6b8-40cd-9633-c5865547d638 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.269407] env[62368]: DEBUG nova.compute.provider_tree [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.378803] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.602323] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.602478] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.602684] env[62368]: DEBUG nova.network.neutron [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.605810] env[62368]: DEBUG nova.objects.instance [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lazy-loading 'flavor' on Instance uuid 1d5a93f0-1068-4514-b01c-011e95f48498 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.650294] env[62368]: DEBUG oslo_vmware.api [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199047, 'name': ReconfigVM_Task, 'duration_secs': 0.519689} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.650601] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1025.655662] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b79c67db-c797-47ae-9d5a-ffb682334960 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.670624] env[62368]: DEBUG oslo_vmware.api [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 1025.670624] env[62368]: value = "task-1199049" [ 1025.670624] env[62368]: _type = "Task" [ 1025.670624] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.678857] env[62368]: DEBUG oslo_vmware.api [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199049, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.774388] env[62368]: DEBUG nova.scheduler.client.report [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.003959] env[62368]: DEBUG nova.compute.manager [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received event network-vif-plugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.003959] env[62368]: DEBUG oslo_concurrency.lockutils [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.003959] env[62368]: DEBUG oslo_concurrency.lockutils [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] Lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.003959] env[62368]: DEBUG oslo_concurrency.lockutils [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] Lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.004122] env[62368]: DEBUG nova.compute.manager [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] No waiting events found dispatching network-vif-plugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1026.004290] env[62368]: WARNING nova.compute.manager [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received unexpected event network-vif-plugged-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 for instance with vm_state shelved_offloaded and task_state spawning. [ 1026.004469] env[62368]: DEBUG nova.compute.manager [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received event network-changed-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.004628] env[62368]: DEBUG nova.compute.manager [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Refreshing instance network info cache due to event network-changed-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1026.004797] env[62368]: DEBUG oslo_concurrency.lockutils [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] Acquiring lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.179967] env[62368]: DEBUG oslo_vmware.api [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199049, 'name': ReconfigVM_Task, 'duration_secs': 0.155033} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.180277] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259876', 'volume_id': 'cf7750c2-86cd-43ff-89ac-880138be3dc5', 'name': 'volume-cf7750c2-86cd-43ff-89ac-880138be3dc5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e09be20-ce79-4bfa-bf84-2877f4e534e6', 'attached_at': '', 'detached_at': '', 'volume_id': 'cf7750c2-86cd-43ff-89ac-880138be3dc5', 'serial': 'cf7750c2-86cd-43ff-89ac-880138be3dc5'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1026.352471] env[62368]: DEBUG nova.network.neutron [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [{"id": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "address": "fa:16:3e:a8:8e:88", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15eed45d-e0", "ovs_interfaceid": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.378908] env[62368]: DEBUG nova.compute.manager [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Stashing vm_state: active {{(pid=62368) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1026.528666] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.528897] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.529061] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1026.612440] env[62368]: DEBUG oslo_concurrency.lockutils [None req-647b29d4-49fe-4217-a6d0-4105e26d469f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.260s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.724377] env[62368]: DEBUG nova.objects.instance [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lazy-loading 'flavor' on Instance uuid 3e09be20-ce79-4bfa-bf84-2877f4e534e6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.785162] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.766s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.790997] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.522s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.791232] env[62368]: DEBUG nova.objects.instance [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1026.855588] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.858121] env[62368]: DEBUG oslo_concurrency.lockutils [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] Acquired lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.858403] env[62368]: DEBUG nova.network.neutron [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Refreshing network info cache for port 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1026.883423] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b2c5fa4ac297ef73a0ef62b015d503eb',container_format='bare',created_at=2024-10-08T23:20:50Z,direct_url=,disk_format='vmdk',id=3b257738-fa51-43dd-83ad-cb3309fa62c0,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1886498350-shelved',owner='6b7091fbcb2f4879b78075ed06b95c0d',properties=ImageMetaProps,protected=,size=31660032,status='active',tags=,updated_at=2024-10-08T23:21:03Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.883698] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.883874] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.884097] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.884262] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.884429] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.884641] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.884829] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.885025] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.885217] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.885411] env[62368]: DEBUG nova.virt.hardware [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.889452] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a17878-ff3e-4572-9c89-83a0cfc219f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.898157] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3195bc-f860-4ba8-b055-34a291b87042 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.912095] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:8e:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15eed45d-e09b-4aac-b3e2-e3b77ed1efe2', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.919800] env[62368]: DEBUG oslo.service.loopingcall [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.920824] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.921097] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1026.921342] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37484b4f-adc1-4540-9a04-3379654b5dd7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.941717] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.941717] env[62368]: value = "task-1199050" [ 1026.941717] env[62368]: _type = "Task" [ 1026.941717] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.949093] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199050, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.454215] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199050, 'name': CreateVM_Task, 'duration_secs': 0.321034} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.454613] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1027.455073] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.455253] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.455639] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.456272] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5b54704-7594-4488-8659-bb3269bd7eec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.461442] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1027.461442] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f48360-6b08-9bd4-bdb8-e43f08d9cc91" [ 1027.461442] env[62368]: _type = "Task" [ 1027.461442] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.471872] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f48360-6b08-9bd4-bdb8-e43f08d9cc91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.636098] env[62368]: DEBUG nova.network.neutron [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updated VIF entry in instance network info cache for port 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1027.636684] env[62368]: DEBUG nova.network.neutron [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [{"id": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "address": "fa:16:3e:a8:8e:88", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15eed45d-e0", "ovs_interfaceid": "15eed45d-e09b-4aac-b3e2-e3b77ed1efe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.696918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "1d5a93f0-1068-4514-b01c-011e95f48498" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.696918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.696918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "1d5a93f0-1068-4514-b01c-011e95f48498-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.696918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.696918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.700597] env[62368]: INFO nova.compute.manager [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Terminating instance [ 1027.702393] env[62368]: DEBUG nova.compute.manager [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1027.702594] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1027.703424] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98adb0d-1730-42e4-a34a-b1afea2fa706 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.711218] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1027.711859] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37216bc1-95f2-4388-be07-697dbf915114 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.718653] env[62368]: DEBUG oslo_vmware.api [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1027.718653] env[62368]: value = "task-1199052" [ 1027.718653] env[62368]: _type = "Task" [ 1027.718653] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.726547] env[62368]: DEBUG oslo_vmware.api [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.734162] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1c0f59cb-3361-42e6-9067-2014610b30d3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.726s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.805877] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b87662a1-a9df-415d-903a-a5c445bf58ff tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.809352] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.431s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.809665] env[62368]: DEBUG nova.objects.instance [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lazy-loading 'resources' on Instance uuid b79f0e79-9e3a-47c7-9949-8743601ec6c5 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.973619] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.973886] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Processing image 3b257738-fa51-43dd-83ad-cb3309fa62c0 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.974178] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0/3b257738-fa51-43dd-83ad-cb3309fa62c0.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.974532] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0/3b257738-fa51-43dd-83ad-cb3309fa62c0.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.974755] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.975046] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d36a27e1-813a-4212-aba6-4a7bb2de7f11 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.986825] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.987025] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1027.987728] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4be61c4-56b9-4de3-9c4a-c15ad467ea8a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.993441] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1027.993441] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f3062c-1891-540d-6b16-490ba3fb89de" [ 1027.993441] env[62368]: _type = "Task" [ 1027.993441] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.001361] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f3062c-1891-540d-6b16-490ba3fb89de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.139588] env[62368]: DEBUG oslo_concurrency.lockutils [req-37efd6c3-6071-49d1-9eec-d8e0eae6340d req-8631849d-fba0-421b-a99b-20a5aee59b11 service nova] Releasing lock "refresh_cache-fa637a2e-a113-4d23-8924-728ba75f202c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.228901] env[62368]: DEBUG oslo_vmware.api [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199052, 'name': PowerOffVM_Task, 'duration_secs': 0.225518} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.229184] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.229357] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1028.229610] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1edf385c-ae6d-4b13-b837-a519f08e8c93 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.314406] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1028.314654] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1028.314842] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleting the datastore file [datastore2] 1d5a93f0-1068-4514-b01c-011e95f48498 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.315125] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fbbe0de-82de-4449-93b0-4ed8b2650e50 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.326967] env[62368]: DEBUG oslo_vmware.api [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1028.326967] env[62368]: value = "task-1199054" [ 1028.326967] env[62368]: _type = "Task" [ 1028.326967] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.331997] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.332225] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.332443] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.332629] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.332808] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.335700] env[62368]: INFO nova.compute.manager [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Swapping old allocation on dict_keys(['2202a74c-753d-4e1d-a031-7cefe24ee9d6']) held by migration 818c9b11-7728-44b9-a2c9-30d7e48036ce for instance [ 1028.340920] env[62368]: INFO nova.compute.manager [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Terminating instance [ 1028.342425] env[62368]: DEBUG oslo_vmware.api [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199054, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.345248] env[62368]: DEBUG nova.compute.manager [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.345444] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1028.346299] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6539038-ebf5-4793-a612-36a2c627953c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.358034] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.358183] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12c8ae0a-e46c-426c-b5a5-629e718b7df5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.360197] env[62368]: DEBUG nova.scheduler.client.report [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Overwriting current allocation {'allocations': {'2202a74c-753d-4e1d-a031-7cefe24ee9d6': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 138}}, 'project_id': '73819709966b4e5f873020b399c49871', 'user_id': '50e00b1a1fd3401fa803dea5b0801965', 'consumer_generation': 1} on consumer e6dd9825-f841-49ac-a75a-83d9de3858ac {{(pid=62368) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1028.366093] env[62368]: DEBUG oslo_vmware.api [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 1028.366093] env[62368]: value = "task-1199055" [ 1028.366093] env[62368]: _type = "Task" [ 1028.366093] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.374173] env[62368]: DEBUG oslo_vmware.api [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199055, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.460218] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.460500] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.460598] env[62368]: DEBUG nova.network.neutron [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.490253] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb2b5e3-aa0b-43c0-a277-ab44bee08643 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.500280] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c42fda8-53af-4e54-81c8-4e4e2e2cef96 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.508506] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Preparing fetch location {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1028.508847] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Fetch image to [datastore1] OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a/OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a.vmdk {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1028.509079] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Downloading stream optimized image 3b257738-fa51-43dd-83ad-cb3309fa62c0 to [datastore1] OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a/OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a.vmdk on the data store datastore1 as vApp {{(pid=62368) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1028.509301] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Downloading image file data 3b257738-fa51-43dd-83ad-cb3309fa62c0 to the ESX as VM named 'OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a' {{(pid=62368) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1028.537875] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77a021c-ed31-4d14-aa8f-9b8d82bb8cb4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.540888] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Didn't find any instances for network info cache update. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1028.541176] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.541728] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.541906] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.542101] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.542275] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.542439] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.542574] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1028.542746] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.547827] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4029ee-2216-4559-88ce-f654ec98a2d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.565687] env[62368]: DEBUG nova.compute.provider_tree [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.622704] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1028.622704] env[62368]: value = "resgroup-9" [ 1028.622704] env[62368]: _type = "ResourcePool" [ 1028.622704] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1028.622983] env[62368]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-eca57ecc-6004-4716-87d8-f439fee04c2a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.644401] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lease: (returnval){ [ 1028.644401] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ceddb9-002c-909f-3e9b-8dd1408d48e3" [ 1028.644401] env[62368]: _type = "HttpNfcLease" [ 1028.644401] env[62368]: } obtained for vApp import into resource pool (val){ [ 1028.644401] env[62368]: value = "resgroup-9" [ 1028.644401] env[62368]: _type = "ResourcePool" [ 1028.644401] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1028.644673] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the lease: (returnval){ [ 1028.644673] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ceddb9-002c-909f-3e9b-8dd1408d48e3" [ 1028.644673] env[62368]: _type = "HttpNfcLease" [ 1028.644673] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1028.653064] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1028.653064] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ceddb9-002c-909f-3e9b-8dd1408d48e3" [ 1028.653064] env[62368]: _type = "HttpNfcLease" [ 1028.653064] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1028.836985] env[62368]: DEBUG oslo_vmware.api [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199054, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224276} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.837267] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.837481] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1028.837672] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1028.837978] env[62368]: INFO nova.compute.manager [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1028.838245] env[62368]: DEBUG oslo.service.loopingcall [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.838462] env[62368]: DEBUG nova.compute.manager [-] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1028.838550] env[62368]: DEBUG nova.network.neutron [-] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1028.878235] env[62368]: DEBUG oslo_vmware.api [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199055, 'name': PowerOffVM_Task, 'duration_secs': 0.189979} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.878496] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.878717] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1028.878968] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b35e478f-60be-45c4-8ebf-1455826f9beb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.942805] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1028.943168] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1028.943449] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleting the datastore file [datastore2] 3e09be20-ce79-4bfa-bf84-2877f4e534e6 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.943710] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4571754c-74b1-4635-b85d-2473576f0fdc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.950133] env[62368]: DEBUG oslo_vmware.api [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 1028.950133] env[62368]: value = "task-1199058" [ 1028.950133] env[62368]: _type = "Task" [ 1028.950133] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.958637] env[62368]: DEBUG oslo_vmware.api [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199058, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.048342] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.068480] env[62368]: DEBUG nova.scheduler.client.report [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.156487] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1029.156487] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ceddb9-002c-909f-3e9b-8dd1408d48e3" [ 1029.156487] env[62368]: _type = "HttpNfcLease" [ 1029.156487] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1029.192370] env[62368]: DEBUG nova.network.neutron [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [{"id": "05119e53-57b5-4b7d-a056-3bb898352179", "address": "fa:16:3e:3b:b9:de", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05119e53-57", "ovs_interfaceid": "05119e53-57b5-4b7d-a056-3bb898352179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.331325] env[62368]: DEBUG nova.compute.manager [req-bf85cd69-334a-4a7e-8b61-02cc087f3212 req-47532924-9fbb-44cb-961a-111b0e9c1f1b service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Received event network-vif-deleted-d47a86a3-3fd6-48d2-8a80-e6b7f201bd76 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.331548] env[62368]: INFO nova.compute.manager [req-bf85cd69-334a-4a7e-8b61-02cc087f3212 req-47532924-9fbb-44cb-961a-111b0e9c1f1b service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Neutron deleted interface d47a86a3-3fd6-48d2-8a80-e6b7f201bd76; detaching it from the instance and deleting it from the info cache [ 1029.331733] env[62368]: DEBUG nova.network.neutron [req-bf85cd69-334a-4a7e-8b61-02cc087f3212 req-47532924-9fbb-44cb-961a-111b0e9c1f1b service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.464211] env[62368]: DEBUG oslo_vmware.api [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199058, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176224} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.464512] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.464653] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1029.464834] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1029.465013] env[62368]: INFO nova.compute.manager [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1029.465268] env[62368]: DEBUG oslo.service.loopingcall [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.465500] env[62368]: DEBUG nova.compute.manager [-] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.465602] env[62368]: DEBUG nova.network.neutron [-] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.508121] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1029.508355] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259893', 'volume_id': '4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'name': 'volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca1c6fcb-9450-4b59-b8b1-39e88aa61af0', 'attached_at': '', 'detached_at': '', 'volume_id': '4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'serial': '4e212d2b-6fb0-4bbf-9c37-a44e83960938'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1029.509232] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8a0134-ff29-40cd-9afe-d58c440514e4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.525244] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9286a5d-0db4-49a2-9db2-56dde2fc0d11 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.551278] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938/volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.551594] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4039d916-9fad-4c88-b971-01ef57219dc7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.570092] env[62368]: DEBUG oslo_vmware.api [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1029.570092] env[62368]: value = "task-1199059" [ 1029.570092] env[62368]: _type = "Task" [ 1029.570092] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.573866] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.765s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.576042] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.655s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.582924] env[62368]: DEBUG oslo_vmware.api [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199059, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.601582] env[62368]: INFO nova.scheduler.client.report [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Deleted allocations for instance b79f0e79-9e3a-47c7-9949-8743601ec6c5 [ 1029.654662] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1029.654662] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ceddb9-002c-909f-3e9b-8dd1408d48e3" [ 1029.654662] env[62368]: _type = "HttpNfcLease" [ 1029.654662] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1029.654988] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1029.654988] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ceddb9-002c-909f-3e9b-8dd1408d48e3" [ 1029.654988] env[62368]: _type = "HttpNfcLease" [ 1029.654988] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1029.655725] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9670b9-384c-49be-9741-554df0120e26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.665753] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283222b-85c6-1f1d-ae0a-3fa5ca147e00/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1029.665938] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating HTTP connection to write to file with size = 31660032 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283222b-85c6-1f1d-ae0a-3fa5ca147e00/disk-0.vmdk. {{(pid=62368) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1029.727368] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-e6dd9825-f841-49ac-a75a-83d9de3858ac" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.728307] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e11e05-0967-43f1-aee6-f3206b372519 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.740379] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bc64ceb5-6916-4a7a-9200-5dce9b9c6b50 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.744117] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecbb203-32f7-4943-be40-29baf0189904 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.803986] env[62368]: DEBUG nova.network.neutron [-] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.837983] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bdb4a128-47cf-400e-8fa0-0f35baacd50b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.842766] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40da87a4-7682-4a38-8c25-d6270ae36607 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.874038] env[62368]: DEBUG nova.compute.manager [req-bf85cd69-334a-4a7e-8b61-02cc087f3212 req-47532924-9fbb-44cb-961a-111b0e9c1f1b service nova] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Detach interface failed, port_id=d47a86a3-3fd6-48d2-8a80-e6b7f201bd76, reason: Instance 1d5a93f0-1068-4514-b01c-011e95f48498 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1030.083159] env[62368]: DEBUG oslo_vmware.api [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199059, 'name': ReconfigVM_Task, 'duration_secs': 0.432959} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.083549] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfigured VM instance instance-00000063 to attach disk [datastore1] volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938/volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.093087] env[62368]: INFO nova.compute.claims [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.096643] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2d1c728-9ac6-4398-8ece-21b37e6ced30 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.116837] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c1ddc991-bed1-43ee-a339-1f800474775d tempest-ServersTestJSON-662825960 tempest-ServersTestJSON-662825960-project-member] Lock "b79f0e79-9e3a-47c7-9949-8743601ec6c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.656s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.121613] env[62368]: DEBUG oslo_vmware.api [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1030.121613] env[62368]: value = "task-1199060" [ 1030.121613] env[62368]: _type = "Task" [ 1030.121613] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.133162] env[62368]: DEBUG oslo_vmware.api [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199060, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.304521] env[62368]: INFO nova.compute.manager [-] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Took 1.47 seconds to deallocate network for instance. [ 1030.436586] env[62368]: DEBUG nova.network.neutron [-] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.614807] env[62368]: INFO nova.compute.resource_tracker [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating resource usage from migration aaaefbde-0cc3-4807-8a58-ece2d52aaba2 [ 1030.632942] env[62368]: DEBUG oslo_vmware.api [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199060, 'name': ReconfigVM_Task, 'duration_secs': 0.157354} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.635216] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259893', 'volume_id': '4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'name': 'volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca1c6fcb-9450-4b59-b8b1-39e88aa61af0', 'attached_at': '', 'detached_at': '', 'volume_id': '4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'serial': '4e212d2b-6fb0-4bbf-9c37-a44e83960938'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1030.782631] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac03036-ebc6-4b5d-b4fd-47aaf935ae76 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.791166] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a49129-1b32-494f-90fc-e0422eae2166 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.823083] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.824270] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3ab709-a8b8-4891-8a02-00acb9a352a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.833457] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a860bbc-1cf8-4c7d-aa61-b4b5740bba8a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.848957] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1030.849702] env[62368]: DEBUG nova.compute.provider_tree [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.855133] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c37f3f79-75cc-42a6-8789-9b3b64834b7b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.863331] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1030.863331] env[62368]: value = "task-1199061" [ 1030.863331] env[62368]: _type = "Task" [ 1030.863331] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.881680] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.942132] env[62368]: INFO nova.compute.manager [-] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Took 1.48 seconds to deallocate network for instance. [ 1031.030019] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Completed reading data from the image iterator. {{(pid=62368) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1031.030265] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283222b-85c6-1f1d-ae0a-3fa5ca147e00/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1031.031185] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3b2096-5d0c-47b8-bf78-58c8f68f19c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.040869] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283222b-85c6-1f1d-ae0a-3fa5ca147e00/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1031.040869] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283222b-85c6-1f1d-ae0a-3fa5ca147e00/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1031.040869] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4c950880-ce83-47da-afd4-08136791ea69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.275500] env[62368]: DEBUG oslo_vmware.rw_handles [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5283222b-85c6-1f1d-ae0a-3fa5ca147e00/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1031.275755] env[62368]: INFO nova.virt.vmwareapi.images [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Downloaded image file data 3b257738-fa51-43dd-83ad-cb3309fa62c0 [ 1031.276721] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b194916-c7fd-47ff-815d-51589ee9d67c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.292723] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5c42854-449c-4da2-9cd4-2fd34e0f581f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.322117] env[62368]: INFO nova.virt.vmwareapi.images [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] The imported VM was unregistered [ 1031.324890] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Caching image {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1031.325172] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating directory with path [datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.325466] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-792fde4a-804a-4b2c-8f2f-581d7228d2de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.337539] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Created directory with path [datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.337539] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a/OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a.vmdk to [datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0/3b257738-fa51-43dd-83ad-cb3309fa62c0.vmdk. {{(pid=62368) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1031.338029] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-1c7e8173-2683-4cf0-b0ab-5d49a5aac089 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.344505] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1031.344505] env[62368]: value = "task-1199063" [ 1031.344505] env[62368]: _type = "Task" [ 1031.344505] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.355287] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199063, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.359410] env[62368]: DEBUG nova.scheduler.client.report [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.374224] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199061, 'name': PowerOffVM_Task, 'duration_secs': 0.294357} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.374450] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1031.375275] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.375479] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.375655] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.375855] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.376018] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.376181] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.376588] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.376686] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.377666] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.378258] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.378473] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.384395] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53f41920-61e6-420b-94a1-ac71d0569372 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.397554] env[62368]: DEBUG nova.compute.manager [req-60896b11-5a83-40f1-97d1-43324eec0b4a req-1d8f416e-c755-4985-b928-5570f044251e service nova] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Received event network-vif-deleted-561cb794-4c95-416d-825a-193df8ddc3ca {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.402916] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1031.402916] env[62368]: value = "task-1199064" [ 1031.402916] env[62368]: _type = "Task" [ 1031.402916] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.412492] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199064, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.453089] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.682308] env[62368]: DEBUG nova.objects.instance [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid ca1c6fcb-9450-4b59-b8b1-39e88aa61af0 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.854868] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199063, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.863986] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.288s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.864597] env[62368]: INFO nova.compute.manager [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Migrating [ 1031.872415] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.824s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.872630] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.872771] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1031.873110] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.050s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.873325] env[62368]: DEBUG nova.objects.instance [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lazy-loading 'resources' on Instance uuid 1d5a93f0-1068-4514-b01c-011e95f48498 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.879299] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee9e978-9d4b-4e6b-ad4e-f1d571c3e120 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.899724] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b120185-f02e-4249-b90f-aa15593beb58 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.919060] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199064, 'name': ReconfigVM_Task, 'duration_secs': 0.407718} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.929533] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d743f9-5941-4ede-bccf-e606183478c6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.933266] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e53f7c-370d-4fd0-9748-f80392963ae9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.959250] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a7f927-d832-48c1-bc96-a04592137e97 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.964343] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.964477] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.964613] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.964840] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.965030] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.965212] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.965429] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.965620] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.965811] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.966020] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.966238] env[62368]: DEBUG nova.virt.hardware [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.967627] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa9a6246-be3c-460d-8a12-ae0d7c785305 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.975310] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1031.975310] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f97f3e-8539-b4ca-e69a-4fe2d89f7e8e" [ 1031.975310] env[62368]: _type = "Task" [ 1031.975310] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.003674] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179201MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1032.003827] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.015435] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f97f3e-8539-b4ca-e69a-4fe2d89f7e8e, 'name': SearchDatastore_Task, 'duration_secs': 0.012673} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.022996] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1032.023401] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e1c5ef4-26bf-4226-aeba-308347ba31a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.047943] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1032.047943] env[62368]: value = "task-1199065" [ 1032.047943] env[62368]: _type = "Task" [ 1032.047943] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.058349] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199065, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.188228] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b69208b2-025c-4de3-8f9b-f37aec0b5421 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.282s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.357149] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199063, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.391031] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.391362] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.391362] env[62368]: DEBUG nova.network.neutron [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1032.538035] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9045d0df-6d7a-413a-ab14-86fb434d2643 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.547064] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fa699a-2cda-4667-b569-2a6e5980cafb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.560042] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199065, 'name': ReconfigVM_Task, 'duration_secs': 0.294091} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.584660] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1032.587247] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fe471a-522c-4233-a4a4-bcac53eee968 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.589381] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f112d4c-31eb-44a9-abbc-d8a699c20d66 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.610978] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e1d87b-9ba1-49a1-9bdc-d45ef99ca1a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.622278] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac/e6dd9825-f841-49ac-a75a-83d9de3858ac.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.622703] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff9f0df4-9576-457a-91c0-2e23c019a379 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.650442] env[62368]: DEBUG nova.compute.provider_tree [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.653127] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1032.653127] env[62368]: value = "task-1199066" [ 1032.653127] env[62368]: _type = "Task" [ 1032.653127] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.663221] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199066, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.856152] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199063, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.109044] env[62368]: DEBUG nova.network.neutron [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance_info_cache with network_info: [{"id": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "address": "fa:16:3e:5b:ce:6e", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8702b61c-a9", "ovs_interfaceid": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.155099] env[62368]: DEBUG nova.scheduler.client.report [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.169538] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199066, 'name': ReconfigVM_Task, 'duration_secs': 0.52299} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.169873] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfigured VM instance instance-0000005c to attach disk [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac/e6dd9825-f841-49ac-a75a-83d9de3858ac.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.170786] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14bdd804-a8f6-412a-9290-cbb4a77fd647 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.198211] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed8f91b-05bc-4db9-a831-951cb42d8087 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.227311] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df83911a-3e8e-4b08-b75a-535dc7bef2d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.254412] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8098bf-b79d-488a-a38e-623618cd9659 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.261944] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1033.262249] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd5cbd35-96bd-4b90-87f4-f68e3898070b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.269187] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1033.269187] env[62368]: value = "task-1199067" [ 1033.269187] env[62368]: _type = "Task" [ 1033.269187] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.279830] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.358973] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199063, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.430025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.430025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.613559] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.665176] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.790s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.666074] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.213s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.666567] env[62368]: DEBUG nova.objects.instance [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lazy-loading 'resources' on Instance uuid 3e09be20-ce79-4bfa-bf84-2877f4e534e6 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.693025] env[62368]: INFO nova.scheduler.client.report [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted allocations for instance 1d5a93f0-1068-4514-b01c-011e95f48498 [ 1033.780786] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199067, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.857426] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199063, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.933606] env[62368]: DEBUG nova.compute.utils [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1034.202461] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77cb28ec-9372-48d3-91a8-63ca7d35d345 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "1d5a93f0-1068-4514-b01c-011e95f48498" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.506s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.280349] env[62368]: DEBUG oslo_vmware.api [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199067, 'name': PowerOnVM_Task, 'duration_secs': 0.711547} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.283485] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1034.338905] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fcc580-1323-4921-a65c-650581fe26a5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.348246] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c93afec-2554-4119-8080-52ace8f9bebc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.359506] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199063, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.929231} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.385332] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a/OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a.vmdk to [datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0/3b257738-fa51-43dd-83ad-cb3309fa62c0.vmdk. [ 1034.385532] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Cleaning up location [datastore1] OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1034.385757] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_1ca4ab7a-b31e-4d24-8155-1ffb6e3a2e3a {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.386672] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-944db1eb-b9ae-4fbd-ba3c-244fb20956f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.389106] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0eba4f9-b15a-4527-8c1b-4e14b4825859 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.399021] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a22aab-2456-44c7-922d-9fd26a0a1661 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.401873] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1034.401873] env[62368]: value = "task-1199068" [ 1034.401873] env[62368]: _type = "Task" [ 1034.401873] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.412704] env[62368]: DEBUG nova.compute.provider_tree [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.419053] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199068, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.437156] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.912692] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199068, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29612} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.912692] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.912692] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0/3b257738-fa51-43dd-83ad-cb3309fa62c0.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.913156] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0/3b257738-fa51-43dd-83ad-cb3309fa62c0.vmdk to [datastore1] fa637a2e-a113-4d23-8924-728ba75f202c/fa637a2e-a113-4d23-8924-728ba75f202c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1034.913238] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9898dd57-7232-484b-a334-ed43164a62e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.915705] env[62368]: DEBUG nova.scheduler.client.report [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.924281] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1034.924281] env[62368]: value = "task-1199069" [ 1034.924281] env[62368]: _type = "Task" [ 1034.924281] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.932593] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199069, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.128197] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089f72b5-81e5-44fc-a25b-f4bd056b49b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.152183] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance 'c3dde3bb-380f-40e8-bc39-ffdba6ba3756' progress to 0 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1035.337295] env[62368]: INFO nova.compute.manager [None req-b7f627e2-0632-4a4b-98bf-7f8b7dd60897 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance to original state: 'active' [ 1035.420871] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.423127] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.419s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.434160] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199069, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.448391] env[62368]: INFO nova.scheduler.client.report [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleted allocations for instance 3e09be20-ce79-4bfa-bf84-2877f4e534e6 [ 1035.505737] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.506020] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.506263] env[62368]: INFO nova.compute.manager [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Attaching volume caa1618e-7c7b-4799-9d5c-23a837766e2f to /dev/sdc [ 1035.541031] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551b65b0-ed22-41ce-b834-0f51ad117874 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.548368] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ee487d-1fe4-4cc9-9f77-eb87380ec71a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.561897] env[62368]: DEBUG nova.virt.block_device [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Updating existing volume attachment record: 4369530b-a69a-4844-8315-577b37209071 {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1035.659466] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1035.659734] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1facb5ef-84fb-4b93-a03d-f1628351aab9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.670267] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1035.670267] env[62368]: value = "task-1199070" [ 1035.670267] env[62368]: _type = "Task" [ 1035.670267] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.680707] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199070, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.946132] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199069, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.956232] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ad7b619b-250c-4e1d-86b3-f7618e8f1ec3 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "3e09be20-ce79-4bfa-bf84-2877f4e534e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.624s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.186161] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199070, 'name': PowerOffVM_Task, 'duration_secs': 0.307468} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.186498] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1036.186696] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance 'c3dde3bb-380f-40e8-bc39-ffdba6ba3756' progress to 17 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1036.203864] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "f5a33613-790b-4324-9b88-1c39f7c4269c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.204243] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "f5a33613-790b-4324-9b88-1c39f7c4269c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.438924] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Applying migration context for instance c3dde3bb-380f-40e8-bc39-ffdba6ba3756 as it has an incoming, in-progress migration aaaefbde-0cc3-4807-8a58-ece2d52aaba2. Migration status is migrating {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1036.439928] env[62368]: INFO nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating resource usage from migration aaaefbde-0cc3-4807-8a58-ece2d52aaba2 [ 1036.447578] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199069, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.469374] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 2a46c954-449e-4d62-be80-add1040ed4c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.469522] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b87094da-6258-469e-ab37-5557955ad3a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.469646] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 69ddb565-6c79-44e5-a7d1-d339ab426fae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.469938] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance ca1c6fcb-9450-4b59-b8b1-39e88aa61af0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.469938] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance e6dd9825-f841-49ac-a75a-83d9de3858ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.470052] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance fa637a2e-a113-4d23-8924-728ba75f202c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.470108] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Migration aaaefbde-0cc3-4807-8a58-ece2d52aaba2 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1036.470223] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance c3dde3bb-380f-40e8-bc39-ffdba6ba3756 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.693999] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1036.694361] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1036.694571] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.694769] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1036.694919] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.695813] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1036.695813] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1036.695813] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1036.695813] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1036.695813] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1036.696036] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.702061] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-411cb606-5ebd-45a4-b36d-db55663945a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.712047] env[62368]: DEBUG nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1036.721947] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1036.721947] env[62368]: value = "task-1199073" [ 1036.721947] env[62368]: _type = "Task" [ 1036.721947] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.734251] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199073, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.945272] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199069, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.973832] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance f5a33613-790b-4324-9b88-1c39f7c4269c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1036.974230] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1036.974278] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1037.052583] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b4070763-a2f2-480f-909c-e323f15a9ac5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.052868] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.124546] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1f35e2-53ea-4c92-b0e4-764fd90e2d53 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.134666] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef7de0d-5279-4aa1-8b22-c93743b93c70 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.169272] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1462a13a-ec38-4b19-a6f8-1ebe44142aff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.178966] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffae854-d2fa-4fb2-89db-77d7c31cb180 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.195911] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.233538] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199073, 'name': ReconfigVM_Task, 'duration_secs': 0.289439} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.234692] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.235126] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance 'c3dde3bb-380f-40e8-bc39-ffdba6ba3756' progress to 33 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1037.323639] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.324051] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.324360] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.324579] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.324760] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.327567] env[62368]: INFO nova.compute.manager [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Terminating instance [ 1037.329684] env[62368]: DEBUG nova.compute.manager [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.329904] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1037.330166] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-292e04f5-90a4-4930-a24c-1cf2ee0e133a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.340850] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1037.340850] env[62368]: value = "task-1199074" [ 1037.340850] env[62368]: _type = "Task" [ 1037.340850] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.350847] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199074, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.445264] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199069, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.554935] env[62368]: DEBUG nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1037.699729] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.743641] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1037.743641] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1037.744188] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1037.744188] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1037.744366] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1037.744530] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1037.744751] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1037.746033] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1037.746033] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1037.746033] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1037.746033] env[62368]: DEBUG nova.virt.hardware [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1037.751423] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1037.752248] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6d03cfe-4f29-4df2-a04f-f4625136347e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.776916] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1037.776916] env[62368]: value = "task-1199075" [ 1037.776916] env[62368]: _type = "Task" [ 1037.776916] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.786040] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199075, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.850695] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199074, 'name': PowerOffVM_Task, 'duration_secs': 0.319478} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.850987] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1037.851223] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1037.851460] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259881', 'volume_id': 'c77431e9-8136-407c-9491-6f6caa1d9991', 'name': 'volume-c77431e9-8136-407c-9491-6f6caa1d9991', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'e6dd9825-f841-49ac-a75a-83d9de3858ac', 'attached_at': '2024-10-08T23:21:30.000000', 'detached_at': '', 'volume_id': 'c77431e9-8136-407c-9491-6f6caa1d9991', 'serial': 'c77431e9-8136-407c-9491-6f6caa1d9991'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1037.852248] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303bd2fa-6206-4ae0-87a3-36bac39c04fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.875712] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255d8bf8-6abf-4560-a453-d0da55104d94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.882991] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06d71fa-d850-4e0c-9104-0ae9541fe681 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.904853] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58372fc8-530c-43ab-a740-ceea50d63d38 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.920907] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] The volume has not been displaced from its original location: [datastore2] volume-c77431e9-8136-407c-9491-6f6caa1d9991/volume-c77431e9-8136-407c-9491-6f6caa1d9991.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1037.926362] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1037.926697] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1c97279-1758-400d-a465-809475c8b242 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.947182] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199069, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.920523} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.948443] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3b257738-fa51-43dd-83ad-cb3309fa62c0/3b257738-fa51-43dd-83ad-cb3309fa62c0.vmdk to [datastore1] fa637a2e-a113-4d23-8924-728ba75f202c/fa637a2e-a113-4d23-8924-728ba75f202c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1037.948812] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1037.948812] env[62368]: value = "task-1199077" [ 1037.948812] env[62368]: _type = "Task" [ 1037.948812] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.949580] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc76bdb-a17e-451d-af97-3e62a86f4937 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.959389] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199077, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.976257] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] fa637a2e-a113-4d23-8924-728ba75f202c/fa637a2e-a113-4d23-8924-728ba75f202c.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.976670] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-757dbff1-ceae-4997-bde1-d4182aba29c3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.998185] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1037.998185] env[62368]: value = "task-1199078" [ 1037.998185] env[62368]: _type = "Task" [ 1037.998185] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.006832] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199078, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.077330] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.205083] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1038.205243] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.782s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.205463] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.971s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.207087] env[62368]: INFO nova.compute.claims [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.285131] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199075, 'name': ReconfigVM_Task, 'duration_secs': 0.193749} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.285464] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1038.286244] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c4bf19-043d-49d2-9a9f-b53061e802c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.315527] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] c3dde3bb-380f-40e8-bc39-ffdba6ba3756/c3dde3bb-380f-40e8-bc39-ffdba6ba3756.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.315853] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-187b3be9-35e3-4870-8be4-a837266bdc1e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.334286] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1038.334286] env[62368]: value = "task-1199079" [ 1038.334286] env[62368]: _type = "Task" [ 1038.334286] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.344021] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199079, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.461899] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199077, 'name': ReconfigVM_Task, 'duration_secs': 0.21654} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.462143] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1038.466744] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57825507-9905-4558-be96-eaac39abf7c7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.482017] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1038.482017] env[62368]: value = "task-1199080" [ 1038.482017] env[62368]: _type = "Task" [ 1038.482017] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.489779] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.507821] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199078, 'name': ReconfigVM_Task, 'duration_secs': 0.318501} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.508095] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Reconfigured VM instance instance-0000005e to attach disk [datastore1] fa637a2e-a113-4d23-8924-728ba75f202c/fa637a2e-a113-4d23-8924-728ba75f202c.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.509203] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'encryption_format': None, 'encrypted': False, 'size': 0, 'device_name': '/dev/sda', 'boot_index': 0, 'disk_bus': None, 'encryption_options': None, 'device_type': 'disk', 'encryption_secret_uuid': None, 'image_id': 'd6bd1b2e-f5f9-49c5-bb6f-9536d441b77d'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'guest_format': None, 'attachment_id': '07dc05ea-dcbe-470c-8dd1-b566875120dd', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259891', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'name': 'volume-8880c5c4-9ad5-4209-884a-84eb03713365', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fa637a2e-a113-4d23-8924-728ba75f202c', 'attached_at': '', 'detached_at': '', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'serial': '8880c5c4-9ad5-4209-884a-84eb03713365'}, 'mount_device': '/dev/sdb', 'boot_index': None, 'disk_bus': None, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=62368) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1038.509411] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1038.509626] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259891', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'name': 'volume-8880c5c4-9ad5-4209-884a-84eb03713365', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fa637a2e-a113-4d23-8924-728ba75f202c', 'attached_at': '', 'detached_at': '', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'serial': '8880c5c4-9ad5-4209-884a-84eb03713365'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1038.510374] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab8fea0-ecc5-4ed5-b741-82e246193ac4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.525184] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71dbe8c-bf16-4f55-96e0-2d2199db5698 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.548406] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] volume-8880c5c4-9ad5-4209-884a-84eb03713365/volume-8880c5c4-9ad5-4209-884a-84eb03713365.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.548668] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f10602e-98e6-4a84-b80e-c9a6f466c8d1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.565603] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1038.565603] env[62368]: value = "task-1199081" [ 1038.565603] env[62368]: _type = "Task" [ 1038.565603] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.572948] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.843549] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199079, 'name': ReconfigVM_Task, 'duration_secs': 0.30271} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.844085] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Reconfigured VM instance instance-00000068 to attach disk [datastore2] c3dde3bb-380f-40e8-bc39-ffdba6ba3756/c3dde3bb-380f-40e8-bc39-ffdba6ba3756.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.844523] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance 'c3dde3bb-380f-40e8-bc39-ffdba6ba3756' progress to 50 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1038.991200] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199080, 'name': ReconfigVM_Task, 'duration_secs': 0.181408} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.991566] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259881', 'volume_id': 'c77431e9-8136-407c-9491-6f6caa1d9991', 'name': 'volume-c77431e9-8136-407c-9491-6f6caa1d9991', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'e6dd9825-f841-49ac-a75a-83d9de3858ac', 'attached_at': '2024-10-08T23:21:30.000000', 'detached_at': '', 'volume_id': 'c77431e9-8136-407c-9491-6f6caa1d9991', 'serial': 'c77431e9-8136-407c-9491-6f6caa1d9991'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1038.991757] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1038.992529] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d7663d-5df1-4cda-a042-0229487566f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.998707] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1038.998935] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2da627c4-b254-4f77-89f0-0e002bd823b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.061271] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1039.061497] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1039.061682] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleting the datastore file [datastore2] e6dd9825-f841-49ac-a75a-83d9de3858ac {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1039.061955] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55bc2b59-ed7d-4531-8769-45cb88b6f07d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.070357] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1039.070357] env[62368]: value = "task-1199083" [ 1039.070357] env[62368]: _type = "Task" [ 1039.070357] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.077462] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199081, 'name': ReconfigVM_Task, 'duration_secs': 0.332491} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.078039] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Reconfigured VM instance instance-0000005e to attach disk [datastore2] volume-8880c5c4-9ad5-4209-884a-84eb03713365/volume-8880c5c4-9ad5-4209-884a-84eb03713365.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.085046] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e90cd54-ecde-4179-a027-5aa0f4206f83 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.094431] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199083, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.099726] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1039.099726] env[62368]: value = "task-1199084" [ 1039.099726] env[62368]: _type = "Task" [ 1039.099726] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.107634] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199084, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.341772] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2c6652-bd84-4ceb-a305-96639606fe06 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.348967] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f1de4f-2e2a-4892-bc28-0cf3b9c43eca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.353883] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2b41f7-332f-4f9c-a4c7-14e872a7a331 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.394606] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ce594c-8e26-435a-8cb8-229b85b9798c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.397847] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e49bbb-bdaa-484d-9ab9-febd557bd517 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.414655] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance 'c3dde3bb-380f-40e8-bc39-ffdba6ba3756' progress to 67 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1039.420903] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb705cd-96d8-40c0-93a0-615a95096995 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.435486] env[62368]: DEBUG nova.compute.provider_tree [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.580896] env[62368]: DEBUG oslo_vmware.api [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199083, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215753} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.583026] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.583026] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1039.583026] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1039.583026] env[62368]: INFO nova.compute.manager [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Took 2.25 seconds to destroy the instance on the hypervisor. [ 1039.583026] env[62368]: DEBUG oslo.service.loopingcall [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.583026] env[62368]: DEBUG nova.compute.manager [-] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.583026] env[62368]: DEBUG nova.network.neutron [-] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1039.608588] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199084, 'name': ReconfigVM_Task, 'duration_secs': 0.204513} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.609192] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259891', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'name': 'volume-8880c5c4-9ad5-4209-884a-84eb03713365', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fa637a2e-a113-4d23-8924-728ba75f202c', 'attached_at': '', 'detached_at': '', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'serial': '8880c5c4-9ad5-4209-884a-84eb03713365'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1039.609525] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c81c496d-1fb7-450d-86b1-8c4d107c7f7f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.617124] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1039.617124] env[62368]: value = "task-1199085" [ 1039.617124] env[62368]: _type = "Task" [ 1039.617124] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.624700] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199085, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.941031] env[62368]: DEBUG nova.scheduler.client.report [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.965500] env[62368]: DEBUG nova.network.neutron [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Port 8702b61c-a9f0-44cc-85e2-3024bb421ef2 binding to destination host cpu-1 is already ACTIVE {{(pid=62368) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1040.081667] env[62368]: DEBUG nova.compute.manager [req-f5a7dfb2-841f-460d-ac94-95c8c7cbb60c req-72534407-e1aa-40da-a755-27684a6adc47 service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Received event network-vif-deleted-05119e53-57b5-4b7d-a056-3bb898352179 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.081667] env[62368]: INFO nova.compute.manager [req-f5a7dfb2-841f-460d-ac94-95c8c7cbb60c req-72534407-e1aa-40da-a755-27684a6adc47 service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Neutron deleted interface 05119e53-57b5-4b7d-a056-3bb898352179; detaching it from the instance and deleting it from the info cache [ 1040.081667] env[62368]: DEBUG nova.network.neutron [req-f5a7dfb2-841f-460d-ac94-95c8c7cbb60c req-72534407-e1aa-40da-a755-27684a6adc47 service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.108338] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1040.108782] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259896', 'volume_id': 'caa1618e-7c7b-4799-9d5c-23a837766e2f', 'name': 'volume-caa1618e-7c7b-4799-9d5c-23a837766e2f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca1c6fcb-9450-4b59-b8b1-39e88aa61af0', 'attached_at': '', 'detached_at': '', 'volume_id': 'caa1618e-7c7b-4799-9d5c-23a837766e2f', 'serial': 'caa1618e-7c7b-4799-9d5c-23a837766e2f'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1040.109796] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad7f522-864e-4a17-8ae0-bc1d9cc4aa03 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.129289] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7fac75-8da7-4011-9601-c9e8cddfc58d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.136596] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199085, 'name': Rename_Task, 'duration_secs': 0.140499} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.153090] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1040.160706] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] volume-caa1618e-7c7b-4799-9d5c-23a837766e2f/volume-caa1618e-7c7b-4799-9d5c-23a837766e2f.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.160854] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ab469c0-ea66-4d9c-8c1b-67c10da7f828 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.162367] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ed54c49-9f40-4802-a04f-f000e4aae675 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.180740] env[62368]: DEBUG oslo_vmware.api [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1040.180740] env[62368]: value = "task-1199087" [ 1040.180740] env[62368]: _type = "Task" [ 1040.180740] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.181987] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1040.181987] env[62368]: value = "task-1199086" [ 1040.181987] env[62368]: _type = "Task" [ 1040.181987] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.192699] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.196388] env[62368]: DEBUG oslo_vmware.api [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199087, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.354103] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "69ddb565-6c79-44e5-a7d1-d339ab426fae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.354432] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "69ddb565-6c79-44e5-a7d1-d339ab426fae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.354682] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "69ddb565-6c79-44e5-a7d1-d339ab426fae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.354879] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "69ddb565-6c79-44e5-a7d1-d339ab426fae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.355092] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "69ddb565-6c79-44e5-a7d1-d339ab426fae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.357365] env[62368]: INFO nova.compute.manager [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Terminating instance [ 1040.359287] env[62368]: DEBUG nova.compute.manager [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.359493] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.360348] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b5f114-4a55-407f-b22a-40054268253b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.368334] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1040.368578] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06ddac28-d380-4d4c-b518-c7673983d37f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.374920] env[62368]: DEBUG oslo_vmware.api [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 1040.374920] env[62368]: value = "task-1199088" [ 1040.374920] env[62368]: _type = "Task" [ 1040.374920] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.444693] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.239s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.445240] env[62368]: DEBUG nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1040.448188] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.371s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.450427] env[62368]: INFO nova.compute.claims [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1040.560849] env[62368]: DEBUG nova.network.neutron [-] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.584230] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb13151c-8240-4cdf-a3f1-afb45b34250d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.594028] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756e91a1-8247-4874-9225-48ab743bd6d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.624985] env[62368]: DEBUG nova.compute.manager [req-f5a7dfb2-841f-460d-ac94-95c8c7cbb60c req-72534407-e1aa-40da-a755-27684a6adc47 service nova] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Detach interface failed, port_id=05119e53-57b5-4b7d-a056-3bb898352179, reason: Instance e6dd9825-f841-49ac-a75a-83d9de3858ac could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1040.693795] env[62368]: DEBUG oslo_vmware.api [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199087, 'name': ReconfigVM_Task, 'duration_secs': 0.429642} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.696520] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfigured VM instance instance-00000063 to attach disk [datastore2] volume-caa1618e-7c7b-4799-9d5c-23a837766e2f/volume-caa1618e-7c7b-4799-9d5c-23a837766e2f.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.701277] env[62368]: DEBUG oslo_vmware.api [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199086, 'name': PowerOnVM_Task, 'duration_secs': 0.495604} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.701477] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ec819d1-73ca-430d-907f-bf5ec8834f69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.711256] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1040.718226] env[62368]: DEBUG oslo_vmware.api [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1040.718226] env[62368]: value = "task-1199089" [ 1040.718226] env[62368]: _type = "Task" [ 1040.718226] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.728597] env[62368]: DEBUG oslo_vmware.api [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199089, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.810713] env[62368]: DEBUG nova.compute.manager [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.811734] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dce89e-7759-4a4e-b74b-679caae019e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.883923] env[62368]: DEBUG oslo_vmware.api [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199088, 'name': PowerOffVM_Task, 'duration_secs': 0.25253} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.884420] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1040.884420] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.884685] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5248eca5-7fcb-4411-8fce-b305e3d6d0b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.945506] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.945729] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.945918] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleting the datastore file [datastore1] 69ddb565-6c79-44e5-a7d1-d339ab426fae {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.946218] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fa8bb8f-5677-46db-bb64-6dadf1f03470 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.953408] env[62368]: DEBUG oslo_vmware.api [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 1040.953408] env[62368]: value = "task-1199091" [ 1040.953408] env[62368]: _type = "Task" [ 1040.953408] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.957399] env[62368]: DEBUG nova.compute.utils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.960623] env[62368]: DEBUG nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1040.960786] env[62368]: DEBUG nova.network.neutron [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1040.967042] env[62368]: DEBUG oslo_vmware.api [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.983549] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.983831] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.984036] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.003834] env[62368]: DEBUG nova.policy [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87c366f937e2485c8b0dc317d6acb8f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '430e72f3b08a4207a2fb4465ff160714', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1041.063285] env[62368]: INFO nova.compute.manager [-] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Took 1.48 seconds to deallocate network for instance. [ 1041.228985] env[62368]: DEBUG oslo_vmware.api [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199089, 'name': ReconfigVM_Task, 'duration_secs': 0.138586} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.229387] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259896', 'volume_id': 'caa1618e-7c7b-4799-9d5c-23a837766e2f', 'name': 'volume-caa1618e-7c7b-4799-9d5c-23a837766e2f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca1c6fcb-9450-4b59-b8b1-39e88aa61af0', 'attached_at': '', 'detached_at': '', 'volume_id': 'caa1618e-7c7b-4799-9d5c-23a837766e2f', 'serial': 'caa1618e-7c7b-4799-9d5c-23a837766e2f'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1041.312123] env[62368]: DEBUG nova.network.neutron [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Successfully created port: fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1041.330020] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6d694af4-39f6-4fd4-be08-71556c4a162b tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 28.117s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.463901] env[62368]: DEBUG nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1041.466149] env[62368]: DEBUG oslo_vmware.api [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.624734] env[62368]: INFO nova.compute.manager [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Took 0.56 seconds to detach 1 volumes for instance. [ 1041.636469] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7544ee15-0704-4837-8a07-d1b55dd3c681 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.644245] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4a669d-7362-4fdc-8baa-7bd708a02722 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.676855] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-595b08d6-02b3-473d-ac8c-2585e7a64313 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.684985] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac1486a-9739-4518-8831-70c98909d382 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.698605] env[62368]: DEBUG nova.compute.provider_tree [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.963474] env[62368]: DEBUG oslo_vmware.api [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.620072} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.963728] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.963914] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1041.964149] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1041.964280] env[62368]: INFO nova.compute.manager [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Took 1.60 seconds to destroy the instance on the hypervisor. [ 1041.964521] env[62368]: DEBUG oslo.service.loopingcall [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.964714] env[62368]: DEBUG nova.compute.manager [-] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1041.964811] env[62368]: DEBUG nova.network.neutron [-] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1042.018596] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.018792] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.018976] env[62368]: DEBUG nova.network.neutron [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.133791] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.201572] env[62368]: DEBUG nova.scheduler.client.report [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.262915] env[62368]: DEBUG nova.compute.manager [req-ed5464b4-7b87-4ee8-9b81-f896d876cda9 req-c551b6d6-2143-4fb3-85da-f7040adcc1c1 service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Received event network-vif-deleted-cb4b1e4c-15e5-4641-b434-e05e31283e5c {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.263188] env[62368]: INFO nova.compute.manager [req-ed5464b4-7b87-4ee8-9b81-f896d876cda9 req-c551b6d6-2143-4fb3-85da-f7040adcc1c1 service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Neutron deleted interface cb4b1e4c-15e5-4641-b434-e05e31283e5c; detaching it from the instance and deleting it from the info cache [ 1042.263336] env[62368]: DEBUG nova.network.neutron [req-ed5464b4-7b87-4ee8-9b81-f896d876cda9 req-c551b6d6-2143-4fb3-85da-f7040adcc1c1 service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.265562] env[62368]: DEBUG nova.objects.instance [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid ca1c6fcb-9450-4b59-b8b1-39e88aa61af0 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.471306] env[62368]: DEBUG nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1042.496207] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.496524] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.496681] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.496873] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.497036] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.497192] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.497434] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.497685] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.497885] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.498068] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.498252] env[62368]: DEBUG nova.virt.hardware [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.499149] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e57b2c7-cf04-40b8-bc11-6c9c0037f70a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.507188] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa87e17-e948-40a7-ab42-071f8aa6a64e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.706877] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.707447] env[62368]: DEBUG nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.710405] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.577s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.710634] env[62368]: DEBUG nova.objects.instance [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'resources' on Instance uuid e6dd9825-f841-49ac-a75a-83d9de3858ac {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.724125] env[62368]: DEBUG nova.network.neutron [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance_info_cache with network_info: [{"id": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "address": "fa:16:3e:5b:ce:6e", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8702b61c-a9", "ovs_interfaceid": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.739983] env[62368]: DEBUG nova.network.neutron [-] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.770419] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9640cffc-3c68-445f-a2f2-e4786f55b2cc tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.264s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.771225] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3cfa5dbd-16ad-4beb-8062-a7b6bcc24d01 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.782492] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045dd4c5-0079-4208-9475-518edcc385cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.810331] env[62368]: DEBUG nova.compute.manager [req-ed5464b4-7b87-4ee8-9b81-f896d876cda9 req-c551b6d6-2143-4fb3-85da-f7040adcc1c1 service nova] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Detach interface failed, port_id=cb4b1e4c-15e5-4641-b434-e05e31283e5c, reason: Instance 69ddb565-6c79-44e5-a7d1-d339ab426fae could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1042.997258] env[62368]: DEBUG nova.compute.manager [req-dbb13f05-225e-405e-99a7-adfa43f04083 req-579724a9-e98b-4cef-b7f9-48d38e26bbc1 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Received event network-vif-plugged-fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.997845] env[62368]: DEBUG oslo_concurrency.lockutils [req-dbb13f05-225e-405e-99a7-adfa43f04083 req-579724a9-e98b-4cef-b7f9-48d38e26bbc1 service nova] Acquiring lock "f5a33613-790b-4324-9b88-1c39f7c4269c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.998206] env[62368]: DEBUG oslo_concurrency.lockutils [req-dbb13f05-225e-405e-99a7-adfa43f04083 req-579724a9-e98b-4cef-b7f9-48d38e26bbc1 service nova] Lock "f5a33613-790b-4324-9b88-1c39f7c4269c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.998400] env[62368]: DEBUG oslo_concurrency.lockutils [req-dbb13f05-225e-405e-99a7-adfa43f04083 req-579724a9-e98b-4cef-b7f9-48d38e26bbc1 service nova] Lock "f5a33613-790b-4324-9b88-1c39f7c4269c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.998576] env[62368]: DEBUG nova.compute.manager [req-dbb13f05-225e-405e-99a7-adfa43f04083 req-579724a9-e98b-4cef-b7f9-48d38e26bbc1 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] No waiting events found dispatching network-vif-plugged-fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.998799] env[62368]: WARNING nova.compute.manager [req-dbb13f05-225e-405e-99a7-adfa43f04083 req-579724a9-e98b-4cef-b7f9-48d38e26bbc1 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Received unexpected event network-vif-plugged-fcd67d36-e1e8-432b-a562-c666ae0c48bf for instance with vm_state building and task_state spawning. [ 1043.077016] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.077297] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.081062] env[62368]: DEBUG nova.network.neutron [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Successfully updated port: fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1043.214054] env[62368]: DEBUG nova.compute.utils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.215478] env[62368]: DEBUG nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1043.215659] env[62368]: DEBUG nova.network.neutron [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1043.226490] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.242731] env[62368]: INFO nova.compute.manager [-] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Took 1.28 seconds to deallocate network for instance. [ 1043.262294] env[62368]: DEBUG nova.policy [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac2df24b03d147f5a33d67a97bbb5bb2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e41fb7a4acf043df959806368ae94ef9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1043.368012] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7975ec-11b0-435c-b6b0-ebaf9f63a197 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.375777] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb69f7f-89a9-407d-b972-cbaf72da6292 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.405286] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c89b69-dbe3-4a4e-b3db-46e7147c306c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.412632] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7de603-3a74-45dc-bd14-035af02a56a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.426135] env[62368]: DEBUG nova.compute.provider_tree [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.563633] env[62368]: DEBUG nova.network.neutron [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Successfully created port: dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.581232] env[62368]: INFO nova.compute.manager [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Detaching volume 4e212d2b-6fb0-4bbf-9c37-a44e83960938 [ 1043.583670] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.583808] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.583955] env[62368]: DEBUG nova.network.neutron [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.624810] env[62368]: INFO nova.virt.block_device [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Attempting to driver detach volume 4e212d2b-6fb0-4bbf-9c37-a44e83960938 from mountpoint /dev/sdb [ 1043.625070] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1043.625269] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259893', 'volume_id': '4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'name': 'volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca1c6fcb-9450-4b59-b8b1-39e88aa61af0', 'attached_at': '', 'detached_at': '', 'volume_id': '4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'serial': '4e212d2b-6fb0-4bbf-9c37-a44e83960938'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1043.626281] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da433bbd-527f-4764-a742-afd9e777b907 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.651475] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72df95a-b39d-4fa4-b87d-e809dc7fe49a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.661226] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f23d0c-61d2-459e-8279-ac006fd5cdcd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.686157] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00bcd3b-fae3-411b-939b-1c08396279f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.701683] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] The volume has not been displaced from its original location: [datastore1] volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938/volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1043.707014] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1043.707322] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b709f7e-3fc0-463a-be8d-454580af72a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.720502] env[62368]: DEBUG nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.725648] env[62368]: DEBUG oslo_vmware.api [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1043.725648] env[62368]: value = "task-1199092" [ 1043.725648] env[62368]: _type = "Task" [ 1043.725648] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.736356] env[62368]: DEBUG oslo_vmware.api [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199092, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.749815] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b64ce8-9b64-483c-a226-b78b60ef30ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.753224] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.770871] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb4ee0e-aeff-43a3-b21d-d1bbf3eeddeb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.778961] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance 'c3dde3bb-380f-40e8-bc39-ffdba6ba3756' progress to 83 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1043.929353] env[62368]: DEBUG nova.scheduler.client.report [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.115403] env[62368]: DEBUG nova.network.neutron [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1044.239232] env[62368]: DEBUG oslo_vmware.api [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199092, 'name': ReconfigVM_Task, 'duration_secs': 0.283961} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.239398] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1044.244241] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8faf948d-a242-4eb7-883f-1f6cf9fe51c2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.255240] env[62368]: DEBUG nova.network.neutron [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Updating instance_info_cache with network_info: [{"id": "fcd67d36-e1e8-432b-a562-c666ae0c48bf", "address": "fa:16:3e:4d:ef:54", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcd67d36-e1", "ovs_interfaceid": "fcd67d36-e1e8-432b-a562-c666ae0c48bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.261988] env[62368]: DEBUG oslo_vmware.api [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1044.261988] env[62368]: value = "task-1199093" [ 1044.261988] env[62368]: _type = "Task" [ 1044.261988] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.270943] env[62368]: DEBUG oslo_vmware.api [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199093, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.284725] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1044.285045] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d4c7170-6eba-4e95-941f-594ac7ffc45c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.290983] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1044.290983] env[62368]: value = "task-1199094" [ 1044.290983] env[62368]: _type = "Task" [ 1044.290983] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.299356] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199094, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.433998] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.436560] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.683s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.436777] env[62368]: DEBUG nova.objects.instance [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lazy-loading 'resources' on Instance uuid 69ddb565-6c79-44e5-a7d1-d339ab426fae {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.455057] env[62368]: INFO nova.scheduler.client.report [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted allocations for instance e6dd9825-f841-49ac-a75a-83d9de3858ac [ 1044.733857] env[62368]: DEBUG nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.759122] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.759404] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.759569] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.759759] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.759917] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.760089] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.760306] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.760464] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.760637] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.760805] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.760979] env[62368]: DEBUG nova.virt.hardware [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.761460] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.761738] env[62368]: DEBUG nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Instance network_info: |[{"id": "fcd67d36-e1e8-432b-a562-c666ae0c48bf", "address": "fa:16:3e:4d:ef:54", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcd67d36-e1", "ovs_interfaceid": "fcd67d36-e1e8-432b-a562-c666ae0c48bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1044.762872] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778ea3ff-ea1d-4d96-84b8-367b47afedc5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.765647] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:ef:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fcd67d36-e1e8-432b-a562-c666ae0c48bf', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.773395] env[62368]: DEBUG oslo.service.loopingcall [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.773681] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1044.777248] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65354e81-2c71-404c-8b10-f058bdb4dbc9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.797139] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1ef5b4-209b-476f-9296-7e9042e89592 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.804343] env[62368]: DEBUG oslo_vmware.api [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199093, 'name': ReconfigVM_Task, 'duration_secs': 0.143651} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.804533] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.804533] env[62368]: value = "task-1199095" [ 1044.804533] env[62368]: _type = "Task" [ 1044.804533] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.805325] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259893', 'volume_id': '4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'name': 'volume-4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca1c6fcb-9450-4b59-b8b1-39e88aa61af0', 'attached_at': '', 'detached_at': '', 'volume_id': '4e212d2b-6fb0-4bbf-9c37-a44e83960938', 'serial': '4e212d2b-6fb0-4bbf-9c37-a44e83960938'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1044.820677] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199094, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.826668] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199095, 'name': CreateVM_Task} progress is 10%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.963335] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d2199ec-08d1-4d39-ac84-7e0d1c2a0920 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "e6dd9825-f841-49ac-a75a-83d9de3858ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.639s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.030215] env[62368]: DEBUG nova.compute.manager [req-84a23486-4c36-4eb9-a950-a90df9ebe085 req-46012784-f8bc-47a9-84aa-2d27f3e8e17d service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Received event network-changed-fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.030215] env[62368]: DEBUG nova.compute.manager [req-84a23486-4c36-4eb9-a950-a90df9ebe085 req-46012784-f8bc-47a9-84aa-2d27f3e8e17d service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Refreshing instance network info cache due to event network-changed-fcd67d36-e1e8-432b-a562-c666ae0c48bf. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1045.030215] env[62368]: DEBUG oslo_concurrency.lockutils [req-84a23486-4c36-4eb9-a950-a90df9ebe085 req-46012784-f8bc-47a9-84aa-2d27f3e8e17d service nova] Acquiring lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.030215] env[62368]: DEBUG oslo_concurrency.lockutils [req-84a23486-4c36-4eb9-a950-a90df9ebe085 req-46012784-f8bc-47a9-84aa-2d27f3e8e17d service nova] Acquired lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.030442] env[62368]: DEBUG nova.network.neutron [req-84a23486-4c36-4eb9-a950-a90df9ebe085 req-46012784-f8bc-47a9-84aa-2d27f3e8e17d service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Refreshing network info cache for port fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.079128] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1793943-bdf0-48ca-802a-c5b5edaebf9c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.087392] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc582ad-2a3d-4416-8110-59e4d91f5f83 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.121992] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d319905-3c65-4980-a0ce-2c870395e5b5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.132826] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edaa6bce-a732-4705-a19d-563a21e8ec6d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.143499] env[62368]: DEBUG nova.compute.provider_tree [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.211313] env[62368]: DEBUG nova.network.neutron [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Successfully updated port: dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.302507] env[62368]: DEBUG oslo_vmware.api [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199094, 'name': PowerOnVM_Task, 'duration_secs': 0.59824} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.302784] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1045.303037] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b410e6-e0cf-4c45-aad3-0d500e79cffe tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance 'c3dde3bb-380f-40e8-bc39-ffdba6ba3756' progress to 100 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1045.317049] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199095, 'name': CreateVM_Task, 'duration_secs': 0.351164} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.317049] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1045.317049] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.317049] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.317049] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1045.317335] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b537eec-7414-4ac0-8bd6-4a3c13084122 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.321089] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1045.321089] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52815618-b800-9a3d-5eb0-7d065212c475" [ 1045.321089] env[62368]: _type = "Task" [ 1045.321089] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.328447] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52815618-b800-9a3d-5eb0-7d065212c475, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.353177] env[62368]: DEBUG nova.objects.instance [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid ca1c6fcb-9450-4b59-b8b1-39e88aa61af0 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.650285] env[62368]: DEBUG nova.scheduler.client.report [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.714166] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.714274] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.714617] env[62368]: DEBUG nova.network.neutron [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1045.737110] env[62368]: DEBUG nova.network.neutron [req-84a23486-4c36-4eb9-a950-a90df9ebe085 req-46012784-f8bc-47a9-84aa-2d27f3e8e17d service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Updated VIF entry in instance network info cache for port fcd67d36-e1e8-432b-a562-c666ae0c48bf. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1045.737534] env[62368]: DEBUG nova.network.neutron [req-84a23486-4c36-4eb9-a950-a90df9ebe085 req-46012784-f8bc-47a9-84aa-2d27f3e8e17d service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Updating instance_info_cache with network_info: [{"id": "fcd67d36-e1e8-432b-a562-c666ae0c48bf", "address": "fa:16:3e:4d:ef:54", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcd67d36-e1", "ovs_interfaceid": "fcd67d36-e1e8-432b-a562-c666ae0c48bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.834758] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52815618-b800-9a3d-5eb0-7d065212c475, 'name': SearchDatastore_Task, 'duration_secs': 0.011439} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.835113] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.835302] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.835549] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.835692] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.835981] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.836191] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48b64b9f-4a28-4b97-8777-0eedc5fe82ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.848985] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.849192] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1045.850183] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d927348-c285-4c13-80d8-3c07f8e7f954 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.857277] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1045.857277] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5258a320-00ce-e9ab-543c-09aa3e21ccac" [ 1045.857277] env[62368]: _type = "Task" [ 1045.857277] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.869501] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5258a320-00ce-e9ab-543c-09aa3e21ccac, 'name': SearchDatastore_Task, 'duration_secs': 0.009629} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.870446] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7cd0070-017d-4f23-b534-b9c9af757900 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.876180] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1045.876180] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]528f60c0-3e46-8361-ea8f-1f4cc89ea75a" [ 1045.876180] env[62368]: _type = "Task" [ 1045.876180] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.884426] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528f60c0-3e46-8361-ea8f-1f4cc89ea75a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.155063] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.718s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.171859] env[62368]: INFO nova.scheduler.client.report [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleted allocations for instance 69ddb565-6c79-44e5-a7d1-d339ab426fae [ 1046.240279] env[62368]: DEBUG oslo_concurrency.lockutils [req-84a23486-4c36-4eb9-a950-a90df9ebe085 req-46012784-f8bc-47a9-84aa-2d27f3e8e17d service nova] Releasing lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.270540] env[62368]: DEBUG nova.network.neutron [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1046.342557] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.365565] env[62368]: DEBUG oslo_concurrency.lockutils [None req-5203e984-edef-4224-97fd-e76a860815b4 tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.288s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.367657] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.025s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.386952] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]528f60c0-3e46-8361-ea8f-1f4cc89ea75a, 'name': SearchDatastore_Task, 'duration_secs': 0.008632} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.387236] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.387500] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] f5a33613-790b-4324-9b88-1c39f7c4269c/f5a33613-790b-4324-9b88-1c39f7c4269c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1046.387756] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd12a2b2-6f2e-4247-a060-0a31149cf0cf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.395910] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1046.395910] env[62368]: value = "task-1199096" [ 1046.395910] env[62368]: _type = "Task" [ 1046.395910] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.403518] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199096, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.442333] env[62368]: DEBUG nova.network.neutron [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updating instance_info_cache with network_info: [{"id": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "address": "fa:16:3e:81:12:31", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbd133bd-21", "ovs_interfaceid": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.663324] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.663711] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.680975] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d571b49-e065-4b86-b185-2126131104bd tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "69ddb565-6c79-44e5-a7d1-d339ab426fae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.326s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.871019] env[62368]: INFO nova.compute.manager [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Detaching volume caa1618e-7c7b-4799-9d5c-23a837766e2f [ 1046.907827] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199096, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479372} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.908148] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] f5a33613-790b-4324-9b88-1c39f7c4269c/f5a33613-790b-4324-9b88-1c39f7c4269c.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1046.908439] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.908719] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e22f256d-4735-458b-ae34-7975fecb217c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.915177] env[62368]: INFO nova.virt.block_device [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Attempting to driver detach volume caa1618e-7c7b-4799-9d5c-23a837766e2f from mountpoint /dev/sdc [ 1046.915393] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1046.915580] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259896', 'volume_id': 'caa1618e-7c7b-4799-9d5c-23a837766e2f', 'name': 'volume-caa1618e-7c7b-4799-9d5c-23a837766e2f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca1c6fcb-9450-4b59-b8b1-39e88aa61af0', 'attached_at': '', 'detached_at': '', 'volume_id': 'caa1618e-7c7b-4799-9d5c-23a837766e2f', 'serial': 'caa1618e-7c7b-4799-9d5c-23a837766e2f'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1046.916375] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adedbfe-77ab-4768-9de8-d4599301049e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.920609] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1046.920609] env[62368]: value = "task-1199097" [ 1046.920609] env[62368]: _type = "Task" [ 1046.920609] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.941374] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b87fee-6462-4935-a51d-60d31f4a717a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.947377] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.947720] env[62368]: DEBUG nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Instance network_info: |[{"id": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "address": "fa:16:3e:81:12:31", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbd133bd-21", "ovs_interfaceid": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1046.948052] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199097, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.948440] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:12:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbd133bd-2126-425e-9772-cbff9f5b8ff7', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.956050] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Creating folder: Project (e41fb7a4acf043df959806368ae94ef9). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1046.956759] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-91e36a92-8237-4baf-b017-cfe00f88cec7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.961070] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a196b9-e12b-40bc-9cc4-c00875a8e358 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.966605] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Created folder: Project (e41fb7a4acf043df959806368ae94ef9) in parent group-v259706. [ 1046.966794] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Creating folder: Instances. Parent ref: group-v259898. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1046.982545] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5e55de1-9fc5-4c07-9699-bee4aae4a422 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.985022] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd3bfb3-7910-46cc-a9ae-1f6ee05cf7db {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.001644] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] The volume has not been displaced from its original location: [datastore2] volume-caa1618e-7c7b-4799-9d5c-23a837766e2f/volume-caa1618e-7c7b-4799-9d5c-23a837766e2f.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1047.007042] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfiguring VM instance instance-00000063 to detach disk 2002 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1047.008555] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f109100f-0b60-42ad-9ba3-0df8c4f117be {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.021400] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Created folder: Instances in parent group-v259898. [ 1047.021644] env[62368]: DEBUG oslo.service.loopingcall [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.021859] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1047.022497] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-946fb994-1137-4eb8-9800-85c8a2eab2bc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.037917] env[62368]: DEBUG oslo_vmware.api [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1047.037917] env[62368]: value = "task-1199100" [ 1047.037917] env[62368]: _type = "Task" [ 1047.037917] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.042759] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.042759] env[62368]: value = "task-1199101" [ 1047.042759] env[62368]: _type = "Task" [ 1047.042759] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.050683] env[62368]: DEBUG oslo_vmware.api [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199100, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.055673] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199101, 'name': CreateVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.058076] env[62368]: DEBUG nova.compute.manager [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Received event network-vif-plugged-dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.058297] env[62368]: DEBUG oslo_concurrency.lockutils [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] Acquiring lock "b4070763-a2f2-480f-909c-e323f15a9ac5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.058505] env[62368]: DEBUG oslo_concurrency.lockutils [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.058673] env[62368]: DEBUG oslo_concurrency.lockutils [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.058842] env[62368]: DEBUG nova.compute.manager [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] No waiting events found dispatching network-vif-plugged-dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1047.059017] env[62368]: WARNING nova.compute.manager [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Received unexpected event network-vif-plugged-dbd133bd-2126-425e-9772-cbff9f5b8ff7 for instance with vm_state building and task_state spawning. [ 1047.059185] env[62368]: DEBUG nova.compute.manager [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Received event network-changed-dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.059357] env[62368]: DEBUG nova.compute.manager [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Refreshing instance network info cache due to event network-changed-dbd133bd-2126-425e-9772-cbff9f5b8ff7. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1047.059580] env[62368]: DEBUG oslo_concurrency.lockutils [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] Acquiring lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.059735] env[62368]: DEBUG oslo_concurrency.lockutils [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] Acquired lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.059878] env[62368]: DEBUG nova.network.neutron [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Refreshing network info cache for port dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1047.068697] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "b87094da-6258-469e-ab37-5557955ad3a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.068921] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "b87094da-6258-469e-ab37-5557955ad3a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.069150] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "b87094da-6258-469e-ab37-5557955ad3a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.069323] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "b87094da-6258-469e-ab37-5557955ad3a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.069491] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "b87094da-6258-469e-ab37-5557955ad3a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.071628] env[62368]: INFO nova.compute.manager [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Terminating instance [ 1047.073416] env[62368]: DEBUG nova.compute.manager [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1047.073613] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1047.075016] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce8f2e4-d6d7-48a5-9c1c-09f52e451a5d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.084025] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.084025] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29ff8c60-fbd5-4190-849a-d7074505d2b2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.090174] env[62368]: DEBUG oslo_vmware.api [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 1047.090174] env[62368]: value = "task-1199102" [ 1047.090174] env[62368]: _type = "Task" [ 1047.090174] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.101915] env[62368]: DEBUG oslo_vmware.api [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199102, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.166626] env[62368]: DEBUG nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1047.430708] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199097, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077732} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.430971] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.431749] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3d618f-7eb3-4f29-847f-5cf33b003b8a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.453045] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] f5a33613-790b-4324-9b88-1c39f7c4269c/f5a33613-790b-4324-9b88-1c39f7c4269c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.453299] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7735c77-aa68-430c-b529-df052c4cd654 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.472259] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1047.472259] env[62368]: value = "task-1199103" [ 1047.472259] env[62368]: _type = "Task" [ 1047.472259] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.479768] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199103, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.549802] env[62368]: DEBUG oslo_vmware.api [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199100, 'name': ReconfigVM_Task, 'duration_secs': 0.264358} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.550441] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Reconfigured VM instance instance-00000063 to detach disk 2002 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1047.558240] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d429427-1101-4645-9232-51b761ad5101 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.567712] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199101, 'name': CreateVM_Task, 'duration_secs': 0.342342} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.569775] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1047.570744] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.570918] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.571262] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.571794] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a260e141-da13-4c65-97c1-52979e85afcf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.574869] env[62368]: DEBUG oslo_vmware.api [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1047.574869] env[62368]: value = "task-1199104" [ 1047.574869] env[62368]: _type = "Task" [ 1047.574869] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.579015] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1047.579015] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52f5b1c6-51be-bac7-d006-40e9ba08d41d" [ 1047.579015] env[62368]: _type = "Task" [ 1047.579015] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.585602] env[62368]: DEBUG oslo_vmware.api [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199104, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.590394] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f5b1c6-51be-bac7-d006-40e9ba08d41d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.598541] env[62368]: DEBUG oslo_vmware.api [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199102, 'name': PowerOffVM_Task, 'duration_secs': 0.229322} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.598773] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1047.598979] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1047.599293] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5267211-a702-4704-8233-4dcf07928b36 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.691469] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1047.691796] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1047.691989] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleting the datastore file [datastore1] b87094da-6258-469e-ab37-5557955ad3a7 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.693127] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.693362] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.697862] env[62368]: INFO nova.compute.claims [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.700418] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9fac813-6eb0-457a-957c-766087800b81 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.706719] env[62368]: DEBUG oslo_vmware.api [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for the task: (returnval){ [ 1047.706719] env[62368]: value = "task-1199106" [ 1047.706719] env[62368]: _type = "Task" [ 1047.706719] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.714881] env[62368]: DEBUG oslo_vmware.api [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.781594] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.781848] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.782053] env[62368]: DEBUG nova.compute.manager [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Going to confirm migration 6 {{(pid=62368) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1047.790040] env[62368]: DEBUG nova.network.neutron [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updated VIF entry in instance network info cache for port dbd133bd-2126-425e-9772-cbff9f5b8ff7. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1047.790680] env[62368]: DEBUG nova.network.neutron [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updating instance_info_cache with network_info: [{"id": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "address": "fa:16:3e:81:12:31", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbd133bd-21", "ovs_interfaceid": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.982745] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199103, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.085781] env[62368]: DEBUG oslo_vmware.api [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199104, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.091124] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52f5b1c6-51be-bac7-d006-40e9ba08d41d, 'name': SearchDatastore_Task, 'duration_secs': 0.01758} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.091479] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.091709] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.091945] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.092111] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.092296] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.092547] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4496f395-e3d4-4dfc-87fd-d60d1d11e278 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.109806] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.110055] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1048.110730] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7111d906-2610-44d9-ab45-5b3d88c2e1da {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.115773] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1048.115773] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521328f3-fdd0-b19c-e2c2-98ac60a1eb5f" [ 1048.115773] env[62368]: _type = "Task" [ 1048.115773] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.123091] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521328f3-fdd0-b19c-e2c2-98ac60a1eb5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.217452] env[62368]: DEBUG oslo_vmware.api [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.293162] env[62368]: DEBUG oslo_concurrency.lockutils [req-0deb0e51-ca1a-40ca-b470-e3327bb2a5d2 req-d4b77c7d-8fa6-4c86-8738-373ef0f297f8 service nova] Releasing lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.319546] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.319833] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquired lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.320039] env[62368]: DEBUG nova.network.neutron [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1048.320239] env[62368]: DEBUG nova.objects.instance [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'info_cache' on Instance uuid c3dde3bb-380f-40e8-bc39-ffdba6ba3756 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.483116] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199103, 'name': ReconfigVM_Task, 'duration_secs': 0.964944} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.483445] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Reconfigured VM instance instance-00000069 to attach disk [datastore2] f5a33613-790b-4324-9b88-1c39f7c4269c/f5a33613-790b-4324-9b88-1c39f7c4269c.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.484084] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c062c732-e778-4e80-acf2-147e6fc56960 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.490323] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1048.490323] env[62368]: value = "task-1199107" [ 1048.490323] env[62368]: _type = "Task" [ 1048.490323] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.498130] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199107, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.585441] env[62368]: DEBUG oslo_vmware.api [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199104, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.625206] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521328f3-fdd0-b19c-e2c2-98ac60a1eb5f, 'name': SearchDatastore_Task, 'duration_secs': 0.045892} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.625987] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-580e6340-bf2a-4f19-b135-f7836e52fb03 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.630906] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1048.630906] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52bbce72-bb60-2b1a-d52e-9baa0252de66" [ 1048.630906] env[62368]: _type = "Task" [ 1048.630906] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.638215] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52bbce72-bb60-2b1a-d52e-9baa0252de66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.718767] env[62368]: DEBUG oslo_vmware.api [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Task: {'id': task-1199106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.683116} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.719064] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.719241] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1048.719433] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1048.719612] env[62368]: INFO nova.compute.manager [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1048.719957] env[62368]: DEBUG oslo.service.loopingcall [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.720219] env[62368]: DEBUG nova.compute.manager [-] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1048.720317] env[62368]: DEBUG nova.network.neutron [-] [instance: b87094da-6258-469e-ab37-5557955ad3a7] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1048.839704] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3015d009-19ab-4a0c-80cb-1a67e3fad3dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.513379] env[62368]: DEBUG nova.compute.manager [req-8af08d1d-d0e0-4f70-9923-f0110c1b7404 req-116d30e4-f929-4e4c-9219-1588c8c97e57 service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Received event network-vif-deleted-3b605458-f2bc-4c5e-8d9a-44fd80ab88bf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.513592] env[62368]: INFO nova.compute.manager [req-8af08d1d-d0e0-4f70-9923-f0110c1b7404 req-116d30e4-f929-4e4c-9219-1588c8c97e57 service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Neutron deleted interface 3b605458-f2bc-4c5e-8d9a-44fd80ab88bf; detaching it from the instance and deleting it from the info cache [ 1049.513780] env[62368]: DEBUG nova.network.neutron [req-8af08d1d-d0e0-4f70-9923-f0110c1b7404 req-116d30e4-f929-4e4c-9219-1588c8c97e57 service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.526922] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a45f28c-5785-477d-a343-ac4acd4d0380 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.529698] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f16b9bbe-b63d-49fd-bc1d-179b8cbd4e71 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.537400] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52bbce72-bb60-2b1a-d52e-9baa0252de66, 'name': SearchDatastore_Task, 'duration_secs': 0.012419} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.566207] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.566510] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] b4070763-a2f2-480f-909c-e323f15a9ac5/b4070763-a2f2-480f-909c-e323f15a9ac5.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1049.566871] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199107, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.567425] env[62368]: DEBUG oslo_vmware.api [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199104, 'name': ReconfigVM_Task, 'duration_secs': 1.142903} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.569963] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ebd5bcb-06a5-408d-b6a1-5cc003916859 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.574321] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688cf52b-3f12-4537-9b33-5738feb78102 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.584198] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259896', 'volume_id': 'caa1618e-7c7b-4799-9d5c-23a837766e2f', 'name': 'volume-caa1618e-7c7b-4799-9d5c-23a837766e2f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca1c6fcb-9450-4b59-b8b1-39e88aa61af0', 'attached_at': '', 'detached_at': '', 'volume_id': 'caa1618e-7c7b-4799-9d5c-23a837766e2f', 'serial': 'caa1618e-7c7b-4799-9d5c-23a837766e2f'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1049.586574] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7cfca2-d395-4a6a-bac4-b425a202b8c8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.597920] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed81910a-d517-4086-b6eb-85b25c6a8a44 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.600804] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1049.600804] env[62368]: value = "task-1199108" [ 1049.600804] env[62368]: _type = "Task" [ 1049.600804] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.608471] env[62368]: DEBUG nova.compute.manager [req-8af08d1d-d0e0-4f70-9923-f0110c1b7404 req-116d30e4-f929-4e4c-9219-1588c8c97e57 service nova] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Detach interface failed, port_id=3b605458-f2bc-4c5e-8d9a-44fd80ab88bf, reason: Instance b87094da-6258-469e-ab37-5557955ad3a7 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1049.620688] env[62368]: DEBUG nova.compute.provider_tree [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.626610] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199108, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.756510] env[62368]: DEBUG nova.network.neutron [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance_info_cache with network_info: [{"id": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "address": "fa:16:3e:5b:ce:6e", "network": {"id": "078acfdf-e26d-4acb-991e-c1f55b106d66", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1918221342-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b70fcc6664f47d7b55447210851c4cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8702b61c-a9", "ovs_interfaceid": "8702b61c-a9f0-44cc-85e2-3024bb421ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.017616] env[62368]: DEBUG nova.network.neutron [-] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.024722] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199107, 'name': Rename_Task, 'duration_secs': 1.242937} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.025646] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1050.026406] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60aa8160-28b9-4b5e-8370-859718eded2c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.036041] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1050.036041] env[62368]: value = "task-1199109" [ 1050.036041] env[62368]: _type = "Task" [ 1050.036041] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.048151] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.117094] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199108, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504586} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.117094] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] b4070763-a2f2-480f-909c-e323f15a9ac5/b4070763-a2f2-480f-909c-e323f15a9ac5.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1050.117528] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.117528] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0578501f-9a3e-406a-bf4b-96db517e66a8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.124083] env[62368]: DEBUG nova.scheduler.client.report [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.127184] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1050.127184] env[62368]: value = "task-1199110" [ 1050.127184] env[62368]: _type = "Task" [ 1050.127184] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.131663] env[62368]: DEBUG nova.objects.instance [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'flavor' on Instance uuid ca1c6fcb-9450-4b59-b8b1-39e88aa61af0 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.137217] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199110, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.259034] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Releasing lock "refresh_cache-c3dde3bb-380f-40e8-bc39-ffdba6ba3756" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.259359] env[62368]: DEBUG nova.objects.instance [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lazy-loading 'migration_context' on Instance uuid c3dde3bb-380f-40e8-bc39-ffdba6ba3756 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.526189] env[62368]: INFO nova.compute.manager [-] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Took 1.81 seconds to deallocate network for instance. [ 1050.544999] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199109, 'name': PowerOnVM_Task} progress is 79%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.629940] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.936s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.630374] env[62368]: DEBUG nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1050.646769] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199110, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058561} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.647089] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.647886] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd999c5-1d13-4668-8d52-be5b44b138c8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.672983] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] b4070763-a2f2-480f-909c-e323f15a9ac5/b4070763-a2f2-480f-909c-e323f15a9ac5.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.673821] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-feadb549-7d51-4da6-99ad-d80b2d3d8106 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.692776] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1050.692776] env[62368]: value = "task-1199111" [ 1050.692776] env[62368]: _type = "Task" [ 1050.692776] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.702666] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199111, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.761901] env[62368]: DEBUG nova.objects.base [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1050.762835] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e593140-52a1-4a01-81c1-fca64bbe04b2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.782648] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99aecb27-a24e-4e03-bfcc-f33e05ef4638 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.788460] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1050.788460] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]525298aa-ba7d-3eaa-1e68-2e4082ac0003" [ 1050.788460] env[62368]: _type = "Task" [ 1050.788460] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.796776] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525298aa-ba7d-3eaa-1e68-2e4082ac0003, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.033120] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.033466] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.033737] env[62368]: DEBUG nova.objects.instance [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lazy-loading 'resources' on Instance uuid b87094da-6258-469e-ab37-5557955ad3a7 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.045565] env[62368]: DEBUG oslo_vmware.api [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199109, 'name': PowerOnVM_Task, 'duration_secs': 0.656121} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.045814] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1051.046024] env[62368]: INFO nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Took 8.57 seconds to spawn the instance on the hypervisor. [ 1051.046239] env[62368]: DEBUG nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.046999] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88627af1-f479-4d15-95d6-2308dfda1a42 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.143028] env[62368]: DEBUG nova.compute.utils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1051.143965] env[62368]: DEBUG oslo_concurrency.lockutils [None req-77c5cf7b-3d08-4d85-a91f-d80a9f5aea6b tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.776s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.145190] env[62368]: DEBUG nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1051.145365] env[62368]: DEBUG nova.network.neutron [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1051.198451] env[62368]: DEBUG nova.policy [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50e00b1a1fd3401fa803dea5b0801965', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73819709966b4e5f873020b399c49871', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1051.206337] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199111, 'name': ReconfigVM_Task, 'duration_secs': 0.276517} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.206674] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Reconfigured VM instance instance-0000006a to attach disk [datastore2] b4070763-a2f2-480f-909c-e323f15a9ac5/b4070763-a2f2-480f-909c-e323f15a9ac5.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.207378] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c4d47509-a3a9-4d14-99ff-1fca53a30f90 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.214864] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1051.214864] env[62368]: value = "task-1199112" [ 1051.214864] env[62368]: _type = "Task" [ 1051.214864] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.228701] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199112, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.303765] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]525298aa-ba7d-3eaa-1e68-2e4082ac0003, 'name': SearchDatastore_Task, 'duration_secs': 0.0074} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.304068] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.497170] env[62368]: DEBUG nova.network.neutron [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Successfully created port: 559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1051.565491] env[62368]: INFO nova.compute.manager [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Took 14.35 seconds to build instance. [ 1051.649015] env[62368]: DEBUG nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1051.668793] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796e62c6-c44c-45fa-a313-ab47198229ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.676784] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46fa2e1-b735-4e9e-94ed-a318d6ff5a4c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.708433] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95444831-0c6e-408e-9eb1-d74f7de24432 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.719603] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa61478b-98a0-4d9d-88c7-306697c58fac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.728293] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199112, 'name': Rename_Task, 'duration_secs': 0.170183} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.736923] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1051.737449] env[62368]: DEBUG nova.compute.provider_tree [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.738913] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d810f1a-f746-49e1-8f7c-3a8a7c3a5f1f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.746155] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1051.746155] env[62368]: value = "task-1199113" [ 1051.746155] env[62368]: _type = "Task" [ 1051.746155] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.755528] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199113, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.812071] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.812420] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.812510] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.812695] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.812948] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.815630] env[62368]: INFO nova.compute.manager [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Terminating instance [ 1051.817487] env[62368]: DEBUG nova.compute.manager [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.817709] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1051.818538] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df1298b-6c26-4884-b409-5610236665d1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.828208] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.828551] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30d9e3a0-582c-4237-9a12-77e4823ba121 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.836182] env[62368]: DEBUG oslo_vmware.api [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1051.836182] env[62368]: value = "task-1199114" [ 1051.836182] env[62368]: _type = "Task" [ 1051.836182] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.844523] env[62368]: DEBUG oslo_vmware.api [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.067722] env[62368]: DEBUG oslo_concurrency.lockutils [None req-88380392-8736-42c4-9a52-b065874e345b tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "f5a33613-790b-4324-9b88-1c39f7c4269c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.863s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.241056] env[62368]: DEBUG nova.scheduler.client.report [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.256739] env[62368]: DEBUG oslo_vmware.api [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199113, 'name': PowerOnVM_Task, 'duration_secs': 0.456688} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.257830] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1052.258127] env[62368]: INFO nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Took 7.52 seconds to spawn the instance on the hypervisor. [ 1052.258317] env[62368]: DEBUG nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.259154] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16960676-62ed-4889-b37c-71625299eac2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.346501] env[62368]: DEBUG oslo_vmware.api [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199114, 'name': PowerOffVM_Task, 'duration_secs': 0.173276} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.346800] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1052.346976] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1052.347248] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb5d86f3-121e-441b-a9c8-4672f3d5f98f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.420612] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1052.420947] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1052.421241] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Deleting the datastore file [datastore2] ca1c6fcb-9450-4b59-b8b1-39e88aa61af0 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.421621] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e12cd20b-87ab-4b9d-b3d0-5a98e221af82 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.429238] env[62368]: DEBUG oslo_vmware.api [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for the task: (returnval){ [ 1052.429238] env[62368]: value = "task-1199116" [ 1052.429238] env[62368]: _type = "Task" [ 1052.429238] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.440860] env[62368]: DEBUG oslo_vmware.api [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199116, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.655799] env[62368]: DEBUG nova.compute.manager [req-03ed2e12-bf81-49e6-b95c-eb79165c9461 req-ad570db0-8a65-40a5-94a3-a28ea3eb8936 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Received event network-changed-fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.655799] env[62368]: DEBUG nova.compute.manager [req-03ed2e12-bf81-49e6-b95c-eb79165c9461 req-ad570db0-8a65-40a5-94a3-a28ea3eb8936 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Refreshing instance network info cache due to event network-changed-fcd67d36-e1e8-432b-a562-c666ae0c48bf. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1052.655908] env[62368]: DEBUG oslo_concurrency.lockutils [req-03ed2e12-bf81-49e6-b95c-eb79165c9461 req-ad570db0-8a65-40a5-94a3-a28ea3eb8936 service nova] Acquiring lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.656129] env[62368]: DEBUG oslo_concurrency.lockutils [req-03ed2e12-bf81-49e6-b95c-eb79165c9461 req-ad570db0-8a65-40a5-94a3-a28ea3eb8936 service nova] Acquired lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.656343] env[62368]: DEBUG nova.network.neutron [req-03ed2e12-bf81-49e6-b95c-eb79165c9461 req-ad570db0-8a65-40a5-94a3-a28ea3eb8936 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Refreshing network info cache for port fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1052.659125] env[62368]: DEBUG nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1052.684007] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1052.684283] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1052.684445] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1052.684742] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1052.684932] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1052.685195] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1052.685425] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1052.685595] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1052.685765] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1052.685931] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1052.686116] env[62368]: DEBUG nova.virt.hardware [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1052.686984] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b6e618-c60b-42d3-bc04-72982d477781 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.694734] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6af0d7b-9cc3-4415-8123-83a45acfa45c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.744697] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.747722] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.443s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.765266] env[62368]: INFO nova.scheduler.client.report [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Deleted allocations for instance b87094da-6258-469e-ab37-5557955ad3a7 [ 1052.776072] env[62368]: INFO nova.compute.manager [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Took 14.72 seconds to build instance. [ 1052.939607] env[62368]: DEBUG oslo_vmware.api [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Task: {'id': task-1199116, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195194} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.940381] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.943064] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1052.943064] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1052.943064] env[62368]: INFO nova.compute.manager [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1052.943064] env[62368]: DEBUG oslo.service.loopingcall [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.943064] env[62368]: DEBUG nova.compute.manager [-] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.943064] env[62368]: DEBUG nova.network.neutron [-] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1052.958457] env[62368]: DEBUG nova.compute.manager [req-bc0d58ac-5afa-49b9-8ffd-e88d3a56ad12 req-bceddc24-6899-4189-958e-3854b93b5cdc service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received event network-vif-plugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.958623] env[62368]: DEBUG oslo_concurrency.lockutils [req-bc0d58ac-5afa-49b9-8ffd-e88d3a56ad12 req-bceddc24-6899-4189-958e-3854b93b5cdc service nova] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.958865] env[62368]: DEBUG oslo_concurrency.lockutils [req-bc0d58ac-5afa-49b9-8ffd-e88d3a56ad12 req-bceddc24-6899-4189-958e-3854b93b5cdc service nova] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.959154] env[62368]: DEBUG oslo_concurrency.lockutils [req-bc0d58ac-5afa-49b9-8ffd-e88d3a56ad12 req-bceddc24-6899-4189-958e-3854b93b5cdc service nova] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.959373] env[62368]: DEBUG nova.compute.manager [req-bc0d58ac-5afa-49b9-8ffd-e88d3a56ad12 req-bceddc24-6899-4189-958e-3854b93b5cdc service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] No waiting events found dispatching network-vif-plugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1052.959523] env[62368]: WARNING nova.compute.manager [req-bc0d58ac-5afa-49b9-8ffd-e88d3a56ad12 req-bceddc24-6899-4189-958e-3854b93b5cdc service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received unexpected event network-vif-plugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 for instance with vm_state building and task_state spawning. [ 1053.280663] env[62368]: DEBUG oslo_concurrency.lockutils [None req-52e0912d-61c4-423f-98a4-c60d1fde3dcf tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.227s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.281436] env[62368]: DEBUG oslo_concurrency.lockutils [None req-42693529-3a31-417f-97f1-ddb2d293d6f0 tempest-ServerRescueNegativeTestJSON-1618015230 tempest-ServerRescueNegativeTestJSON-1618015230-project-member] Lock "b87094da-6258-469e-ab37-5557955ad3a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.212s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.377255] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996b071b-fb73-4fe8-9da4-3179cdd5583e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.387980] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088638d4-1735-49d9-a207-034b8d0554d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.389805] env[62368]: DEBUG nova.network.neutron [req-03ed2e12-bf81-49e6-b95c-eb79165c9461 req-ad570db0-8a65-40a5-94a3-a28ea3eb8936 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Updated VIF entry in instance network info cache for port fcd67d36-e1e8-432b-a562-c666ae0c48bf. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1053.390318] env[62368]: DEBUG nova.network.neutron [req-03ed2e12-bf81-49e6-b95c-eb79165c9461 req-ad570db0-8a65-40a5-94a3-a28ea3eb8936 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Updating instance_info_cache with network_info: [{"id": "fcd67d36-e1e8-432b-a562-c666ae0c48bf", "address": "fa:16:3e:4d:ef:54", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcd67d36-e1", "ovs_interfaceid": "fcd67d36-e1e8-432b-a562-c666ae0c48bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.425273] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da000d3-f7c1-44d7-a12c-5c553abc2ced {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.433985] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36d63e7-63b2-4d59-802a-4962c6c47ec6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.438952] env[62368]: DEBUG nova.network.neutron [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Successfully updated port: 559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1053.452413] env[62368]: DEBUG nova.compute.provider_tree [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.713836] env[62368]: DEBUG nova.compute.manager [req-00e9b6c7-9734-4a9b-bada-e3db6c80e63a req-ebf3db23-7d3f-4014-9cc0-acacae228683 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received event network-changed-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.713962] env[62368]: DEBUG nova.compute.manager [req-00e9b6c7-9734-4a9b-bada-e3db6c80e63a req-ebf3db23-7d3f-4014-9cc0-acacae228683 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Refreshing instance network info cache due to event network-changed-559f003f-8c7a-488b-ae15-6cfa36d5ead5. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1053.714587] env[62368]: DEBUG oslo_concurrency.lockutils [req-00e9b6c7-9734-4a9b-bada-e3db6c80e63a req-ebf3db23-7d3f-4014-9cc0-acacae228683 service nova] Acquiring lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.714765] env[62368]: DEBUG oslo_concurrency.lockutils [req-00e9b6c7-9734-4a9b-bada-e3db6c80e63a req-ebf3db23-7d3f-4014-9cc0-acacae228683 service nova] Acquired lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.714958] env[62368]: DEBUG nova.network.neutron [req-00e9b6c7-9734-4a9b-bada-e3db6c80e63a req-ebf3db23-7d3f-4014-9cc0-acacae228683 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Refreshing network info cache for port 559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1053.893270] env[62368]: DEBUG oslo_concurrency.lockutils [req-03ed2e12-bf81-49e6-b95c-eb79165c9461 req-ad570db0-8a65-40a5-94a3-a28ea3eb8936 service nova] Releasing lock "refresh_cache-f5a33613-790b-4324-9b88-1c39f7c4269c" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.941519] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.959588] env[62368]: DEBUG nova.scheduler.client.report [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.181866] env[62368]: DEBUG nova.network.neutron [-] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.305883] env[62368]: DEBUG nova.network.neutron [req-00e9b6c7-9734-4a9b-bada-e3db6c80e63a req-ebf3db23-7d3f-4014-9cc0-acacae228683 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1054.448616] env[62368]: DEBUG nova.network.neutron [req-00e9b6c7-9734-4a9b-bada-e3db6c80e63a req-ebf3db23-7d3f-4014-9cc0-acacae228683 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.685789] env[62368]: INFO nova.compute.manager [-] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Took 1.74 seconds to deallocate network for instance. [ 1054.687016] env[62368]: DEBUG nova.compute.manager [req-6bfad58d-cd6a-4464-8478-08e938072b9e req-8cc26196-7d63-4d57-9a7f-8a1c17f88cb1 service nova] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Received event network-vif-deleted-76aee038-386b-401a-8c49-5ee204436ab3 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.950735] env[62368]: DEBUG oslo_concurrency.lockutils [req-00e9b6c7-9734-4a9b-bada-e3db6c80e63a req-ebf3db23-7d3f-4014-9cc0-acacae228683 service nova] Releasing lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.951162] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.951328] env[62368]: DEBUG nova.network.neutron [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1054.969920] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.223s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.195554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.195864] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.196168] env[62368]: DEBUG nova.objects.instance [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lazy-loading 'resources' on Instance uuid ca1c6fcb-9450-4b59-b8b1-39e88aa61af0 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.490396] env[62368]: DEBUG nova.network.neutron [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1055.527459] env[62368]: INFO nova.scheduler.client.report [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted allocation for migration aaaefbde-0cc3-4807-8a58-ece2d52aaba2 [ 1055.633973] env[62368]: DEBUG nova.network.neutron [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [{"id": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "address": "fa:16:3e:58:06:cd", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f003f-8c", "ovs_interfaceid": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.740782] env[62368]: DEBUG nova.compute.manager [req-b37f660a-6953-4344-b89c-24fc301755b3 req-c4fdfebe-165b-4030-af9c-cbd42ad288b6 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Received event network-changed-dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.740987] env[62368]: DEBUG nova.compute.manager [req-b37f660a-6953-4344-b89c-24fc301755b3 req-c4fdfebe-165b-4030-af9c-cbd42ad288b6 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Refreshing instance network info cache due to event network-changed-dbd133bd-2126-425e-9772-cbff9f5b8ff7. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1055.741233] env[62368]: DEBUG oslo_concurrency.lockutils [req-b37f660a-6953-4344-b89c-24fc301755b3 req-c4fdfebe-165b-4030-af9c-cbd42ad288b6 service nova] Acquiring lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.741385] env[62368]: DEBUG oslo_concurrency.lockutils [req-b37f660a-6953-4344-b89c-24fc301755b3 req-c4fdfebe-165b-4030-af9c-cbd42ad288b6 service nova] Acquired lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.741551] env[62368]: DEBUG nova.network.neutron [req-b37f660a-6953-4344-b89c-24fc301755b3 req-c4fdfebe-165b-4030-af9c-cbd42ad288b6 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Refreshing network info cache for port dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1055.796995] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf393a6-6a70-4989-899e-ca6404147b3d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.806787] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31079e59-a175-4958-b2e6-6d562ca950e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.837241] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871b25a3-a9cb-4173-9bd1-03e52fa614f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.844233] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65640703-b14f-4a91-a5b8-262ad7217ad5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.857183] env[62368]: DEBUG nova.compute.provider_tree [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.035517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.253s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.136765] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.137153] env[62368]: DEBUG nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Instance network_info: |[{"id": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "address": "fa:16:3e:58:06:cd", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f003f-8c", "ovs_interfaceid": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1056.137636] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:06:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0685bd0b-3dbf-4a06-951c-c6a4726dd4b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '559f003f-8c7a-488b-ae15-6cfa36d5ead5', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.145561] env[62368]: DEBUG oslo.service.loopingcall [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.145786] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1056.147044] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8dccc16-300f-4252-9685-1c2963356fbd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.166446] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.166446] env[62368]: value = "task-1199118" [ 1056.166446] env[62368]: _type = "Task" [ 1056.166446] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.173982] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199118, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.359870] env[62368]: DEBUG nova.scheduler.client.report [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.546388] env[62368]: DEBUG nova.network.neutron [req-b37f660a-6953-4344-b89c-24fc301755b3 req-c4fdfebe-165b-4030-af9c-cbd42ad288b6 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updated VIF entry in instance network info cache for port dbd133bd-2126-425e-9772-cbff9f5b8ff7. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1056.546388] env[62368]: DEBUG nova.network.neutron [req-b37f660a-6953-4344-b89c-24fc301755b3 req-c4fdfebe-165b-4030-af9c-cbd42ad288b6 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updating instance_info_cache with network_info: [{"id": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "address": "fa:16:3e:81:12:31", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbd133bd-21", "ovs_interfaceid": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.678770] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199118, 'name': CreateVM_Task, 'duration_secs': 0.433874} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.678976] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1056.679856] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.680047] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.680380] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1056.680635] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3634b328-76e2-4ca7-9a34-b17789f6cdec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.684870] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1056.684870] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52eb0a1b-7030-57da-cfca-cf7059cefedc" [ 1056.684870] env[62368]: _type = "Task" [ 1056.684870] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.692743] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52eb0a1b-7030-57da-cfca-cf7059cefedc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.808343] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.808631] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.868112] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.672s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.890225] env[62368]: INFO nova.scheduler.client.report [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Deleted allocations for instance ca1c6fcb-9450-4b59-b8b1-39e88aa61af0 [ 1057.048753] env[62368]: DEBUG oslo_concurrency.lockutils [req-b37f660a-6953-4344-b89c-24fc301755b3 req-c4fdfebe-165b-4030-af9c-cbd42ad288b6 service nova] Releasing lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.195946] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52eb0a1b-7030-57da-cfca-cf7059cefedc, 'name': SearchDatastore_Task, 'duration_secs': 0.032394} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.196375] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.196689] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1057.197011] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.197209] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.197452] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1057.197762] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-950b80a0-eb69-4e75-86e9-ce8dcf17f801 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.209257] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.209528] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.211238] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1057.211425] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1057.212771] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da45c800-ffe4-4aff-8587-c81ab3f039b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.219243] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1057.219243] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52b0debe-8761-1475-e3da-1a2e099dcf9d" [ 1057.219243] env[62368]: _type = "Task" [ 1057.219243] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.229746] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b0debe-8761-1475-e3da-1a2e099dcf9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.310726] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.311064] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.311318] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.311554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.311797] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.313521] env[62368]: DEBUG nova.compute.manager [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1057.316136] env[62368]: INFO nova.compute.manager [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Terminating instance [ 1057.317959] env[62368]: DEBUG nova.compute.manager [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1057.318171] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1057.318995] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04aa9d03-caed-4a31-8db7-df92554b311f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.326381] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1057.326660] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71c0cd76-509e-48b4-ac60-d6da40e6c317 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.332038] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1057.332038] env[62368]: value = "task-1199120" [ 1057.332038] env[62368]: _type = "Task" [ 1057.332038] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.339350] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.400567] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a441ec7e-cc3a-4d1a-973b-2bbdddb21edb tempest-AttachVolumeTestJSON-686678081 tempest-AttachVolumeTestJSON-686678081-project-member] Lock "ca1c6fcb-9450-4b59-b8b1-39e88aa61af0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.588s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.713878] env[62368]: DEBUG nova.compute.manager [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1057.734286] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52b0debe-8761-1475-e3da-1a2e099dcf9d, 'name': SearchDatastore_Task, 'duration_secs': 0.011277} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.735588] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74ef848e-1b3d-43ad-b056-215b58270006 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.741450] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1057.741450] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52c464f6-59b3-a671-fb6b-456e38452ad3" [ 1057.741450] env[62368]: _type = "Task" [ 1057.741450] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.749844] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c464f6-59b3-a671-fb6b-456e38452ad3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.838248] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.838519] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.840071] env[62368]: INFO nova.compute.claims [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.845365] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199120, 'name': PowerOffVM_Task, 'duration_secs': 0.316336} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.845833] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1057.846022] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1057.846274] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f0f3a99-8343-4de4-ac70-c969bce224f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.944875] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1057.945120] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1057.945355] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleting the datastore file [datastore2] c3dde3bb-380f-40e8-bc39-ffdba6ba3756 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.945638] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93e62a4f-315e-402a-a573-70433b2848c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.953465] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for the task: (returnval){ [ 1057.953465] env[62368]: value = "task-1199123" [ 1057.953465] env[62368]: _type = "Task" [ 1057.953465] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.961417] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199123, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.232709] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.253046] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52c464f6-59b3-a671-fb6b-456e38452ad3, 'name': SearchDatastore_Task, 'duration_secs': 0.064622} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.253330] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.253601] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2/a1fc23d4-dcbb-4df9-bd5a-74377bee36d2.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1058.253861] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29737283-3848-4567-a10d-aa3e673a4418 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.261302] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1058.261302] env[62368]: value = "task-1199124" [ 1058.261302] env[62368]: _type = "Task" [ 1058.261302] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.268589] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.463344] env[62368]: DEBUG oslo_vmware.api [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Task: {'id': task-1199123, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.491741} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.463606] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.463792] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1058.463973] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1058.464167] env[62368]: INFO nova.compute.manager [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1058.464412] env[62368]: DEBUG oslo.service.loopingcall [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.464646] env[62368]: DEBUG nova.compute.manager [-] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1058.464739] env[62368]: DEBUG nova.network.neutron [-] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1058.768827] env[62368]: DEBUG nova.compute.manager [req-99ad1825-18be-40f4-bdca-1e3add949005 req-ab191c2a-6dff-4ba5-90f9-5ddd6c8a40a7 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Received event network-vif-deleted-8702b61c-a9f0-44cc-85e2-3024bb421ef2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.769301] env[62368]: INFO nova.compute.manager [req-99ad1825-18be-40f4-bdca-1e3add949005 req-ab191c2a-6dff-4ba5-90f9-5ddd6c8a40a7 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Neutron deleted interface 8702b61c-a9f0-44cc-85e2-3024bb421ef2; detaching it from the instance and deleting it from the info cache [ 1058.769593] env[62368]: DEBUG nova.network.neutron [req-99ad1825-18be-40f4-bdca-1e3add949005 req-ab191c2a-6dff-4ba5-90f9-5ddd6c8a40a7 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.776988] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199124, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.975063] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3dabc48-4885-48d1-89bd-e93fc580cf4a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.984812] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9bc98f-31ae-4d88-9458-2e47723f88ee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.019594] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9729ed9-a417-4848-8ee5-bd74a14dfd5e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.028018] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8684a31d-c6c1-4ab0-b49d-f151801daf94 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.042862] env[62368]: DEBUG nova.compute.provider_tree [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.247857] env[62368]: DEBUG nova.network.neutron [-] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.272674] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199124, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.837683} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.272916] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b78f3d3b-7bd7-49f3-a644-cf912648ff9f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.274876] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2/a1fc23d4-dcbb-4df9-bd5a-74377bee36d2.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1059.275104] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1059.275557] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bfdff1c5-5719-4809-bcc0-3c627b66de68 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.285356] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33146f53-c4bf-4e7d-8db9-b28a3a544a34 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.296237] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1059.296237] env[62368]: value = "task-1199126" [ 1059.296237] env[62368]: _type = "Task" [ 1059.296237] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.304405] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199126, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.313156] env[62368]: DEBUG nova.compute.manager [req-99ad1825-18be-40f4-bdca-1e3add949005 req-ab191c2a-6dff-4ba5-90f9-5ddd6c8a40a7 service nova] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Detach interface failed, port_id=8702b61c-a9f0-44cc-85e2-3024bb421ef2, reason: Instance c3dde3bb-380f-40e8-bc39-ffdba6ba3756 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1059.546143] env[62368]: DEBUG nova.scheduler.client.report [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.750416] env[62368]: INFO nova.compute.manager [-] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Took 1.29 seconds to deallocate network for instance. [ 1059.808142] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199126, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070826} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.808305] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1059.809052] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6848ca0-a07f-4b5e-8614-86306b8a5399 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.832274] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2/a1fc23d4-dcbb-4df9-bd5a-74377bee36d2.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.832992] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7ec9154-9f14-413a-a007-f0bb123cb8d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.852874] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1059.852874] env[62368]: value = "task-1199127" [ 1059.852874] env[62368]: _type = "Task" [ 1059.852874] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.861165] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199127, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.052100] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.213s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.053080] env[62368]: DEBUG nova.compute.manager [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1060.057623] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.825s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.060189] env[62368]: INFO nova.compute.claims [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.257741] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.364226] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199127, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.565975] env[62368]: DEBUG nova.compute.utils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1060.570446] env[62368]: DEBUG nova.compute.manager [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Not allocating networking since 'none' was specified. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1060.862922] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.864024] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199127, 'name': ReconfigVM_Task, 'duration_secs': 0.559381} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.864493] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Reconfigured VM instance instance-0000006b to attach disk [datastore1] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2/a1fc23d4-dcbb-4df9-bd5a-74377bee36d2.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.865127] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-44ca97b4-f2e8-415a-bfec-443d5f692644 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.871310] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1060.871310] env[62368]: value = "task-1199129" [ 1060.871310] env[62368]: _type = "Task" [ 1060.871310] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.880334] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199129, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.073884] env[62368]: DEBUG nova.compute.manager [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1061.169616] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a7e064-3768-4f5d-a2be-1912c217f23d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.176460] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0da67c-7324-4c5f-ab18-ea5a46097069 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.207948] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4a7fd9-a3bf-4105-a7e5-46cb082ef83b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.216044] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c63a24-482a-4cc5-8a47-30033974ba28 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.230101] env[62368]: DEBUG nova.compute.provider_tree [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.381509] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199129, 'name': Rename_Task, 'duration_secs': 0.134193} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.381896] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1061.382126] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36f7383d-4817-4162-9810-cce4cf3c2898 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.388307] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1061.388307] env[62368]: value = "task-1199131" [ 1061.388307] env[62368]: _type = "Task" [ 1061.388307] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.395914] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.733858] env[62368]: DEBUG nova.scheduler.client.report [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.899575] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199131, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.082983] env[62368]: DEBUG nova.compute.manager [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1062.109939] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.110237] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.110399] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.110582] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.110734] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.110975] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.111240] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.111414] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.111606] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.111787] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.111991] env[62368]: DEBUG nova.virt.hardware [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.112881] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a2250f-d74b-4306-b507-6c4acf9a5a8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.120832] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1359582d-9f26-4910-a133-c92eb77c5b77 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.135021] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.140594] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Creating folder: Project (df715f013c784117b9a9cb55482b644d). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1062.140886] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3b77dcd-fd10-4ccd-af3c-c47eee2e56ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.151640] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Created folder: Project (df715f013c784117b9a9cb55482b644d) in parent group-v259706. [ 1062.151836] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Creating folder: Instances. Parent ref: group-v259903. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1062.152071] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98cc8e15-7d16-4ca6-9027-d705d2407888 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.161097] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Created folder: Instances in parent group-v259903. [ 1062.161325] env[62368]: DEBUG oslo.service.loopingcall [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.161513] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1062.161707] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6196bf94-a1c0-4d86-b3f6-be2dae2f9335 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.177393] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.177393] env[62368]: value = "task-1199134" [ 1062.177393] env[62368]: _type = "Task" [ 1062.177393] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.184241] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199134, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.238413] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.239027] env[62368]: DEBUG nova.compute.manager [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1062.242249] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.985s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.242513] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.265185] env[62368]: INFO nova.scheduler.client.report [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Deleted allocations for instance c3dde3bb-380f-40e8-bc39-ffdba6ba3756 [ 1062.363083] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.363297] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Cleaning up deleted instances {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1062.399493] env[62368]: DEBUG oslo_vmware.api [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199131, 'name': PowerOnVM_Task, 'duration_secs': 0.624489} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.399859] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1062.399973] env[62368]: INFO nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Took 9.74 seconds to spawn the instance on the hypervisor. [ 1062.400127] env[62368]: DEBUG nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1062.400886] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af1f61a-a23f-493f-9e92-8cd881003195 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.687073] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199134, 'name': CreateVM_Task, 'duration_secs': 0.302411} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.687257] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1062.687737] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.687924] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.688284] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1062.688543] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b79cf64a-b03b-493e-90d0-a3bcfff69051 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.692705] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1062.692705] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529d1266-1ec4-96d3-8f51-3b2fcbd31eff" [ 1062.692705] env[62368]: _type = "Task" [ 1062.692705] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.699793] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529d1266-1ec4-96d3-8f51-3b2fcbd31eff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.746301] env[62368]: DEBUG nova.compute.utils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.747993] env[62368]: DEBUG nova.compute.manager [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Not allocating networking since 'none' was specified. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1062.772911] env[62368]: DEBUG oslo_concurrency.lockutils [None req-977db45a-a0c9-4eb3-b3fd-58f5396050c8 tempest-DeleteServersTestJSON-1232669160 tempest-DeleteServersTestJSON-1232669160-project-member] Lock "c3dde3bb-380f-40e8-bc39-ffdba6ba3756" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.461s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.878856] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] There are 58 instances to clean {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1062.879070] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: c3dde3bb-380f-40e8-bc39-ffdba6ba3756] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1062.917219] env[62368]: INFO nova.compute.manager [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Took 15.24 seconds to build instance. [ 1063.203021] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529d1266-1ec4-96d3-8f51-3b2fcbd31eff, 'name': SearchDatastore_Task, 'duration_secs': 0.011115} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.203354] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.203597] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1063.203875] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.204040] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.204253] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1063.204600] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0e1c627-bb3e-43f2-ad59-1b2b38bccae5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.212762] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1063.212968] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1063.213729] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b502af5f-2eb8-4496-883c-0bbbf1a8c51f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.220840] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1063.220840] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527a2f6c-eb4f-5744-288e-dee6498234b7" [ 1063.220840] env[62368]: _type = "Task" [ 1063.220840] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.229339] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527a2f6c-eb4f-5744-288e-dee6498234b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.249278] env[62368]: DEBUG nova.compute.manager [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1063.381936] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: f409466d-273b-4880-9e45-844909723e4f] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1063.418856] env[62368]: DEBUG oslo_concurrency.lockutils [None req-3a6506ed-c310-4052-9410-c6cc9a9c86e2 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.755s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.719850] env[62368]: DEBUG nova.compute.manager [req-ac8c40c3-c730-4408-a3f6-2238b78267f7 req-c1a87517-74bd-44b5-b76e-3884ba719cfe service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received event network-changed-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.720067] env[62368]: DEBUG nova.compute.manager [req-ac8c40c3-c730-4408-a3f6-2238b78267f7 req-c1a87517-74bd-44b5-b76e-3884ba719cfe service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Refreshing instance network info cache due to event network-changed-559f003f-8c7a-488b-ae15-6cfa36d5ead5. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1063.720415] env[62368]: DEBUG oslo_concurrency.lockutils [req-ac8c40c3-c730-4408-a3f6-2238b78267f7 req-c1a87517-74bd-44b5-b76e-3884ba719cfe service nova] Acquiring lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.720618] env[62368]: DEBUG oslo_concurrency.lockutils [req-ac8c40c3-c730-4408-a3f6-2238b78267f7 req-c1a87517-74bd-44b5-b76e-3884ba719cfe service nova] Acquired lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.720790] env[62368]: DEBUG nova.network.neutron [req-ac8c40c3-c730-4408-a3f6-2238b78267f7 req-c1a87517-74bd-44b5-b76e-3884ba719cfe service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Refreshing network info cache for port 559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1063.733393] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527a2f6c-eb4f-5744-288e-dee6498234b7, 'name': SearchDatastore_Task, 'duration_secs': 0.008423} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.734722] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9956e373-f41c-46d5-a4d6-283fd806cdfc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.740331] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1063.740331] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5259bf29-d21a-a925-1538-fef9cf561f57" [ 1063.740331] env[62368]: _type = "Task" [ 1063.740331] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.749999] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5259bf29-d21a-a925-1538-fef9cf561f57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.885706] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 9c1598f8-9e9d-4abc-9bd6-901e7e6fdde1] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.253040] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5259bf29-d21a-a925-1538-fef9cf561f57, 'name': SearchDatastore_Task, 'duration_secs': 0.009291} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.253040] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.253188] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] eb9281af-c914-4cf1-bc32-4cf43fb98c81/eb9281af-c914-4cf1-bc32-4cf43fb98c81.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1064.253457] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55cfb4c6-a48b-4194-b707-2311564457a4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.258103] env[62368]: DEBUG nova.compute.manager [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1064.261581] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1064.261581] env[62368]: value = "task-1199138" [ 1064.261581] env[62368]: _type = "Task" [ 1064.261581] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.273732] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199138, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.285368] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.285659] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.285830] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.286043] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.286236] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.286398] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.286658] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.286833] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.287026] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.287228] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.287424] env[62368]: DEBUG nova.virt.hardware [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.288416] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddf0a1e-056b-4683-9978-e0d71af2d6ed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.295973] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af3ad20-0b46-4578-bba9-b7e8589524f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.310492] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1064.316373] env[62368]: DEBUG oslo.service.loopingcall [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1064.319017] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1064.319288] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8934534-6649-452f-afeb-0ab9803c7344 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.336188] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1064.336188] env[62368]: value = "task-1199139" [ 1064.336188] env[62368]: _type = "Task" [ 1064.336188] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.343836] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199139, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.388847] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: ffa476fb-fb7f-4f68-a38a-85dc9c009857] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.506609] env[62368]: DEBUG nova.network.neutron [req-ac8c40c3-c730-4408-a3f6-2238b78267f7 req-c1a87517-74bd-44b5-b76e-3884ba719cfe service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updated VIF entry in instance network info cache for port 559f003f-8c7a-488b-ae15-6cfa36d5ead5. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1064.507088] env[62368]: DEBUG nova.network.neutron [req-ac8c40c3-c730-4408-a3f6-2238b78267f7 req-c1a87517-74bd-44b5-b76e-3884ba719cfe service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [{"id": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "address": "fa:16:3e:58:06:cd", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f003f-8c", "ovs_interfaceid": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.772435] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199138, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459748} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.772894] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] eb9281af-c914-4cf1-bc32-4cf43fb98c81/eb9281af-c914-4cf1-bc32-4cf43fb98c81.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1064.773163] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.773434] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7208ec3e-c993-4b70-9c39-a4ac91b31890 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.779540] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1064.779540] env[62368]: value = "task-1199140" [ 1064.779540] env[62368]: _type = "Task" [ 1064.779540] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.788826] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199140, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.846979] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199139, 'name': CreateVM_Task, 'duration_secs': 0.309591} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.847220] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1064.847602] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.847821] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.848171] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1064.848431] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af17b737-25db-4a70-bf75-8ddeb063341b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.852839] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1064.852839] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52405867-1047-3aae-6e21-c72cae31b8ba" [ 1064.852839] env[62368]: _type = "Task" [ 1064.852839] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.860823] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52405867-1047-3aae-6e21-c72cae31b8ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.892401] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 66e96455-fbe8-4a0d-8fe2-9ad60d5abade] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.009983] env[62368]: DEBUG oslo_concurrency.lockutils [req-ac8c40c3-c730-4408-a3f6-2238b78267f7 req-c1a87517-74bd-44b5-b76e-3884ba719cfe service nova] Releasing lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.289414] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199140, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.363097] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52405867-1047-3aae-6e21-c72cae31b8ba, 'name': SearchDatastore_Task, 'duration_secs': 0.009718} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.363464] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.363707] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1065.363949] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.364185] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.364454] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1065.364797] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85b71c37-6b10-40ce-9da5-b3e63474138d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.372749] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1065.372936] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1065.373721] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-361430e2-cb24-44a6-8a6c-c6512ce9ec02 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.378622] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1065.378622] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5210d1a1-1198-ab69-093e-8dfe49114e1a" [ 1065.378622] env[62368]: _type = "Task" [ 1065.378622] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.386775] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5210d1a1-1198-ab69-093e-8dfe49114e1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.396404] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: ca1c6fcb-9450-4b59-b8b1-39e88aa61af0] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.790370] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199140, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.631947} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.790663] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.791676] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6ee9ef-8ead-49bf-a8ee-c71d94f9d7b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.811906] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] eb9281af-c914-4cf1-bc32-4cf43fb98c81/eb9281af-c914-4cf1-bc32-4cf43fb98c81.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.812349] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a25b1f8-16aa-4967-8ff3-d349b4e748bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.832708] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1065.832708] env[62368]: value = "task-1199142" [ 1065.832708] env[62368]: _type = "Task" [ 1065.832708] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.840638] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199142, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.888034] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5210d1a1-1198-ab69-093e-8dfe49114e1a, 'name': SearchDatastore_Task, 'duration_secs': 0.009601} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.888814] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab72d337-28ba-453d-b223-a39cfd1384de {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.894520] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1065.894520] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52ea54a0-a4e3-5540-0ae0-23e5c84e3436" [ 1065.894520] env[62368]: _type = "Task" [ 1065.894520] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.899438] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 84bb7208-670f-4e1d-862a-a480f7db0f42] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.905114] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ea54a0-a4e3-5540-0ae0-23e5c84e3436, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.344775] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199142, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.404920] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52ea54a0-a4e3-5540-0ae0-23e5c84e3436, 'name': SearchDatastore_Task, 'duration_secs': 0.011751} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.405213] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.405482] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3/988cac3a-1637-4682-ba30-e9e1f73cc3e3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1066.405891] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: c1b238e7-3636-47db-b86e-3fcfe917f705] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.407694] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fdf69fe3-7376-420f-bb98-de5c25d4585a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.414841] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1066.414841] env[62368]: value = "task-1199143" [ 1066.414841] env[62368]: _type = "Task" [ 1066.414841] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.422780] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.847613] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199142, 'name': ReconfigVM_Task, 'duration_secs': 0.772059} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.848040] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Reconfigured VM instance instance-0000006c to attach disk [datastore2] eb9281af-c914-4cf1-bc32-4cf43fb98c81/eb9281af-c914-4cf1-bc32-4cf43fb98c81.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.848812] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d15b55f-d6eb-46c5-b653-47f3f7bfc0f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.856777] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1066.856777] env[62368]: value = "task-1199144" [ 1066.856777] env[62368]: _type = "Task" [ 1066.856777] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.868796] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199144, 'name': Rename_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.911677] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: cf3601d5-2956-4301-a7cc-bc2d33bf3790] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.926882] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199143, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459034} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.927204] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3/988cac3a-1637-4682-ba30-e9e1f73cc3e3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1066.927443] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1066.927718] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30a66e3e-28c4-4b68-a657-1ccac4a5375c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.934697] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1066.934697] env[62368]: value = "task-1199145" [ 1066.934697] env[62368]: _type = "Task" [ 1066.934697] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.944726] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199145, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.367507] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199144, 'name': Rename_Task, 'duration_secs': 0.15757} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.369056] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1067.369056] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef6e1faf-bd6f-49b8-aebe-958f4b8886ac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.374844] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1067.374844] env[62368]: value = "task-1199146" [ 1067.374844] env[62368]: _type = "Task" [ 1067.374844] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.383278] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199146, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.417774] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 1d5a93f0-1068-4514-b01c-011e95f48498] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1067.445182] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199145, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06891} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.445529] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1067.446703] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87052cb5-e0c5-4fc9-b3f3-341319c88cf9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.467760] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3/988cac3a-1637-4682-ba30-e9e1f73cc3e3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1067.468397] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ccc40b3-5a24-48df-a14a-f6e460bc9b7c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.489626] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1067.489626] env[62368]: value = "task-1199148" [ 1067.489626] env[62368]: _type = "Task" [ 1067.489626] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.498172] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199148, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.896248] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199146, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.919544] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 72d21aef-0ba0-4f5d-ae43-f5a0e6ce6fd2] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.000013] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.386203] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199146, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.423623] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: e6dd9825-f841-49ac-a75a-83d9de3858ac] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.500102] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.886711] env[62368]: DEBUG oslo_vmware.api [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199146, 'name': PowerOnVM_Task, 'duration_secs': 1.241145} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.886922] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1068.887149] env[62368]: INFO nova.compute.manager [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Took 6.80 seconds to spawn the instance on the hypervisor. [ 1068.887335] env[62368]: DEBUG nova.compute.manager [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.888122] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23c065e-7e66-4cb0-807d-db8ee1bc99d0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.926259] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: e427d283-1f8f-4c7a-869c-4cd48a595470] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.000491] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199148, 'name': ReconfigVM_Task, 'duration_secs': 1.073961} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.000760] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3/988cac3a-1637-4682-ba30-e9e1f73cc3e3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.001394] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eef09eb7-33eb-4519-9872-56282bf07f1d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.007700] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1069.007700] env[62368]: value = "task-1199149" [ 1069.007700] env[62368]: _type = "Task" [ 1069.007700] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.014850] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199149, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.405390] env[62368]: INFO nova.compute.manager [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Took 11.59 seconds to build instance. [ 1069.429514] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 5633292c-9a74-4c2d-893b-6823f4a587a2] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.517271] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199149, 'name': Rename_Task, 'duration_secs': 0.326638} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.517568] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1069.517882] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-906a7d3a-0931-4e4a-b73f-4ea117a9c168 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.523540] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1069.523540] env[62368]: value = "task-1199151" [ 1069.523540] env[62368]: _type = "Task" [ 1069.523540] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.531275] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199151, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.907651] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cc1b19f7-ed96-45c0-85db-458de00f23db tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.099s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.932765] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 3e09be20-ce79-4bfa-bf84-2877f4e534e6] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.033457] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199151, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.445766] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 7a52fab4-f780-4eb2-a1b4-25dfa8caa9dc] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.535711] env[62368]: DEBUG oslo_vmware.api [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199151, 'name': PowerOnVM_Task, 'duration_secs': 0.514417} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.536029] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1070.536240] env[62368]: INFO nova.compute.manager [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Took 6.28 seconds to spawn the instance on the hypervisor. [ 1070.536423] env[62368]: DEBUG nova.compute.manager [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.537228] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99847b4-e688-4e52-bdad-4769fb9f2cef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.949441] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: ad21bf3a-068e-44a7-b58f-2978a631b972] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.052456] env[62368]: INFO nova.compute.manager [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Took 12.83 seconds to build instance. [ 1071.453097] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 14c358a0-434d-472a-a382-b5b0e9bf2bd8] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.554173] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bf5a71da-04a3-4c08-be94-9853112469fe tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.344s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.956368] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a5cbadbd-20dd-4514-8867-20243af5db0c] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.247417] env[62368]: INFO nova.compute.manager [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Rebuilding instance [ 1072.438361] env[62368]: DEBUG nova.compute.manager [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1072.439267] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6568ece2-4fd9-4bb6-85f6-940a1e68c04e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.459085] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 207de202-ca50-4811-84c5-a0d4454d29f7] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.715181] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "753bbfe2-405e-4e81-8756-9f52cb637969" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.716046] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.950975] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1072.951257] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f72fe1c1-3e7b-44a0-a823-b722dea49b3a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.958943] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1072.958943] env[62368]: value = "task-1199152" [ 1072.958943] env[62368]: _type = "Task" [ 1072.958943] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.962144] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 5af32858-dc9a-4380-827a-daf384c29bed] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.971860] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199152, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.217642] env[62368]: DEBUG nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1073.465219] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 55eecf3d-501c-490d-bddd-0211fd082841] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.469927] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199152, 'name': PowerOffVM_Task, 'duration_secs': 0.188867} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.470351] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1073.470570] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1073.471324] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9c81f6-2f8a-46c9-8abe-98069ab8c624 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.477944] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1073.478307] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a0a60bd2-c9fa-4430-8c24-f7d098c98fe1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.504683] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1073.504949] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1073.505156] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Deleting the datastore file [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.505407] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2deb8d7a-fa45-4a73-aac5-7a5edafb91aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.511295] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1073.511295] env[62368]: value = "task-1199154" [ 1073.511295] env[62368]: _type = "Task" [ 1073.511295] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.518745] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.739491] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.739791] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.741342] env[62368]: INFO nova.compute.claims [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1073.971611] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 69ddb565-6c79-44e5-a7d1-d339ab426fae] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.020786] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199154, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14397} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.021059] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.021254] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1074.021435] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1074.475186] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b87094da-6258-469e-ab37-5557955ad3a7] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.849784] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa7ff3e-da13-4b74-98a4-8ceb84d6fc95 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.857315] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc49754-3ba4-42d0-969b-641452fdbacd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.887270] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07976934-2856-4070-b946-027b1a468193 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.894150] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0097acb2-c6b7-472f-a344-66abbc13eb20 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.907125] env[62368]: DEBUG nova.compute.provider_tree [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.978979] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 0cb0097b-2908-4107-bb57-0241915ccff7] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.054674] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1075.054971] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1075.055164] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1075.055358] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1075.055515] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1075.055673] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1075.055889] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1075.056079] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1075.056260] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1075.056430] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1075.056608] env[62368]: DEBUG nova.virt.hardware [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.057520] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e612500a-579a-43d2-b712-7272cfa370b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.065563] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8964c5-09ae-4a05-9767-059e9f0f2008 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.078844] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Instance VIF info [] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.084396] env[62368]: DEBUG oslo.service.loopingcall [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.084629] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1075.084839] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b423f0e-b21a-4ed0-bdf3-6cf9f2a3e9f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.101312] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1075.101312] env[62368]: value = "task-1199155" [ 1075.101312] env[62368]: _type = "Task" [ 1075.101312] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.108661] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199155, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.410278] env[62368]: DEBUG nova.scheduler.client.report [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.482533] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: f397e102-bfbd-4753-b1a4-4c5e9126b2c6] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.612706] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199155, 'name': CreateVM_Task, 'duration_secs': 0.241008} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.612865] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1075.613248] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.613415] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.613744] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1075.613997] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-932a62f4-9d8c-48bf-8eac-3a8937d1abc8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.618329] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1075.618329] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52cf8a52-2ba5-6bda-8e29-6d996032a19c" [ 1075.618329] env[62368]: _type = "Task" [ 1075.618329] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.625414] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cf8a52-2ba5-6bda-8e29-6d996032a19c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.915755] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.176s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.916160] env[62368]: DEBUG nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1075.985556] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 5ed9cee5-8cc8-4b70-96a9-737019363638] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.128699] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cf8a52-2ba5-6bda-8e29-6d996032a19c, 'name': SearchDatastore_Task, 'duration_secs': 0.009281} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.128945] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.129202] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1076.129442] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.129593] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.129789] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1076.130065] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4c62551-5349-400e-a731-e653f2208e43 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.138221] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1076.138399] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1076.139112] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21a438c6-e3ef-4f18-93f7-8e4b6353dd97 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.143849] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1076.143849] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e9b5c3-1dd9-2c40-b4a3-b0900418a8a0" [ 1076.143849] env[62368]: _type = "Task" [ 1076.143849] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.151674] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e9b5c3-1dd9-2c40-b4a3-b0900418a8a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.421175] env[62368]: DEBUG nova.compute.utils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1076.422595] env[62368]: DEBUG nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1076.422775] env[62368]: DEBUG nova.network.neutron [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1076.469716] env[62368]: DEBUG nova.policy [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87c366f937e2485c8b0dc317d6acb8f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '430e72f3b08a4207a2fb4465ff160714', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1076.488824] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 044ac2e4-ddeb-4e8d-bc7f-0bb5474cb4b5] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.655556] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e9b5c3-1dd9-2c40-b4a3-b0900418a8a0, 'name': SearchDatastore_Task, 'duration_secs': 0.009023} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.656434] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37d708b2-81fc-46ab-ad3f-cd4b3bd5bb8d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.662581] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1076.662581] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]520fbd59-f949-5e2d-cff8-0fa1a26ec3c9" [ 1076.662581] env[62368]: _type = "Task" [ 1076.662581] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.669809] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520fbd59-f949-5e2d-cff8-0fa1a26ec3c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.742696] env[62368]: DEBUG nova.network.neutron [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Successfully created port: 11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1076.925890] env[62368]: DEBUG nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1076.991838] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 5554732b-34a1-41bd-9d0d-9bd39d62bec3] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.172898] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]520fbd59-f949-5e2d-cff8-0fa1a26ec3c9, 'name': SearchDatastore_Task, 'duration_secs': 0.008993} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.173228] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.173502] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3/988cac3a-1637-4682-ba30-e9e1f73cc3e3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1077.173829] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a64f107-1f85-4ea1-a8dc-9d54c6b7999b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.180590] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1077.180590] env[62368]: value = "task-1199156" [ 1077.180590] env[62368]: _type = "Task" [ 1077.180590] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.188401] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199156, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.431225] env[62368]: INFO nova.virt.block_device [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Booting with volume a5a81572-db8f-4766-b46a-5e14fa56e64c at /dev/sda [ 1077.472550] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5abcab5c-0052-4bf9-bc75-ed0b47c9a578 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.483309] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e7203e-9ec2-477b-b8f8-68f86825d47a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.495434] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 58efc1cf-2469-41be-bf99-fc7b8c72113c] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.513263] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c02332ba-5a31-4c6c-b0e1-309736ef51b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.522552] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2e7ccb-4477-4b96-8e02-4e16392e68fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.552591] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052712e9-713e-4488-8ca5-0d48d053c2fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.560903] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc70265c-430a-4038-934c-f23e275ba1be {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.575613] env[62368]: DEBUG nova.virt.block_device [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating existing volume attachment record: 3852ace7-918a-4578-9cd0-b4c4ec4a9801 {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1077.692286] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199156, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.946720] env[62368]: DEBUG oslo_concurrency.lockutils [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.947052] env[62368]: DEBUG oslo_concurrency.lockutils [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.999085] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: e6dde32b-46ad-4b4b-8b0e-67e6b7d2145f] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.147398] env[62368]: DEBUG nova.compute.manager [req-1ffdc8d0-27cf-4eae-a786-79bcdb59837c req-a6d26e68-b8a4-4177-9fdc-25f2f5f933bb service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Received event network-vif-plugged-11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1078.147612] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ffdc8d0-27cf-4eae-a786-79bcdb59837c req-a6d26e68-b8a4-4177-9fdc-25f2f5f933bb service nova] Acquiring lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.148054] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ffdc8d0-27cf-4eae-a786-79bcdb59837c req-a6d26e68-b8a4-4177-9fdc-25f2f5f933bb service nova] Lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.148258] env[62368]: DEBUG oslo_concurrency.lockutils [req-1ffdc8d0-27cf-4eae-a786-79bcdb59837c req-a6d26e68-b8a4-4177-9fdc-25f2f5f933bb service nova] Lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.149250] env[62368]: DEBUG nova.compute.manager [req-1ffdc8d0-27cf-4eae-a786-79bcdb59837c req-a6d26e68-b8a4-4177-9fdc-25f2f5f933bb service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] No waiting events found dispatching network-vif-plugged-11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1078.150065] env[62368]: WARNING nova.compute.manager [req-1ffdc8d0-27cf-4eae-a786-79bcdb59837c req-a6d26e68-b8a4-4177-9fdc-25f2f5f933bb service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Received unexpected event network-vif-plugged-11de7a9b-926a-47f0-a28d-a30419725a7f for instance with vm_state building and task_state block_device_mapping. [ 1078.192786] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199156, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.767019} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.193114] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3/988cac3a-1637-4682-ba30-e9e1f73cc3e3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1078.193292] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1078.193552] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-133edcf6-3736-49df-9d01-dde9259ee140 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.200071] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1078.200071] env[62368]: value = "task-1199157" [ 1078.200071] env[62368]: _type = "Task" [ 1078.200071] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.208638] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199157, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.443303] env[62368]: DEBUG nova.network.neutron [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Successfully updated port: 11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1078.450157] env[62368]: INFO nova.compute.manager [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Detaching volume 8880c5c4-9ad5-4209-884a-84eb03713365 [ 1078.490268] env[62368]: INFO nova.virt.block_device [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Attempting to driver detach volume 8880c5c4-9ad5-4209-884a-84eb03713365 from mountpoint /dev/sdb [ 1078.490524] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1078.490744] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259891', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'name': 'volume-8880c5c4-9ad5-4209-884a-84eb03713365', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fa637a2e-a113-4d23-8924-728ba75f202c', 'attached_at': '', 'detached_at': '', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'serial': '8880c5c4-9ad5-4209-884a-84eb03713365'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1078.491657] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9608b0-5907-4481-a39c-2e3982659ede {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.512883] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: cda0cf42-11ae-4da3-b838-5bac4bc2f4d1] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.515447] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833874df-9ee2-4a3d-8346-bdede8495677 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.523348] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca0e41e-9676-40ff-a98f-2f9209e18bcc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.544121] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92310701-fc23-433e-baf3-cfce71e33e3b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.560378] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] The volume has not been displaced from its original location: [datastore2] volume-8880c5c4-9ad5-4209-884a-84eb03713365/volume-8880c5c4-9ad5-4209-884a-84eb03713365.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1078.566281] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1078.566965] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc45532a-f6dd-42fe-aec4-5a0c18438eb3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.586024] env[62368]: DEBUG oslo_vmware.api [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1078.586024] env[62368]: value = "task-1199158" [ 1078.586024] env[62368]: _type = "Task" [ 1078.586024] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.595150] env[62368]: DEBUG oslo_vmware.api [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199158, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.713365] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199157, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077627} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.713647] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1078.714448] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46be6f5a-d238-4f34-851e-a75ccb62f844 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.735882] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3/988cac3a-1637-4682-ba30-e9e1f73cc3e3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.736231] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64948625-4909-4f7d-a308-bddc811b3e5b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.756736] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1078.756736] env[62368]: value = "task-1199159" [ 1078.756736] env[62368]: _type = "Task" [ 1078.756736] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.765165] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199159, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.947468] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.947622] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.947802] env[62368]: DEBUG nova.network.neutron [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1079.019630] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 29d2f0d6-7a76-4039-acbb-a3abd69d8370] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.096935] env[62368]: DEBUG oslo_vmware.api [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199158, 'name': ReconfigVM_Task, 'duration_secs': 0.484017} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.096935] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1079.101460] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8fbc5fc6-ea8f-4a7c-b9b5-fc124816e6f1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.117620] env[62368]: DEBUG oslo_vmware.api [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1079.117620] env[62368]: value = "task-1199160" [ 1079.117620] env[62368]: _type = "Task" [ 1079.117620] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.126110] env[62368]: DEBUG oslo_vmware.api [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199160, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.267251] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199159, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.479731] env[62368]: DEBUG nova.network.neutron [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1079.523679] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: d865d433-2341-4f6c-b840-609530ae2e51] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.614826] env[62368]: DEBUG nova.network.neutron [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance_info_cache with network_info: [{"id": "11de7a9b-926a-47f0-a28d-a30419725a7f", "address": "fa:16:3e:b0:ab:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11de7a9b-92", "ovs_interfaceid": "11de7a9b-926a-47f0-a28d-a30419725a7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.627153] env[62368]: DEBUG oslo_vmware.api [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199160, 'name': ReconfigVM_Task, 'duration_secs': 0.160294} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.628045] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259891', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'name': 'volume-8880c5c4-9ad5-4209-884a-84eb03713365', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fa637a2e-a113-4d23-8924-728ba75f202c', 'attached_at': '', 'detached_at': '', 'volume_id': '8880c5c4-9ad5-4209-884a-84eb03713365', 'serial': '8880c5c4-9ad5-4209-884a-84eb03713365'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1079.667400] env[62368]: DEBUG nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1079.667944] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1079.668187] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1079.668393] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1079.668599] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1079.668754] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1079.668941] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1079.669311] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1079.669545] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1079.669734] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1079.669909] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1079.670103] env[62368]: DEBUG nova.virt.hardware [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1079.670949] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-788a6f3e-4c7a-4a4b-a3bb-88c9ebe5c1e5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.680562] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1890cb8-6d9e-455a-bcd1-b5ae1855de61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.769879] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199159, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.027696] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 0b861a8b-dcad-46ad-90ef-a46450c5cd51] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.118122] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.118469] env[62368]: DEBUG nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Instance network_info: |[{"id": "11de7a9b-926a-47f0-a28d-a30419725a7f", "address": "fa:16:3e:b0:ab:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11de7a9b-92", "ovs_interfaceid": "11de7a9b-926a-47f0-a28d-a30419725a7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1080.118897] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:ab:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '11de7a9b-926a-47f0-a28d-a30419725a7f', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1080.127411] env[62368]: DEBUG oslo.service.loopingcall [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.127735] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1080.128084] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef9ad7b4-4d4b-4a14-b31a-e62a5b89d7ab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.147858] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1080.147858] env[62368]: value = "task-1199161" [ 1080.147858] env[62368]: _type = "Task" [ 1080.147858] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.156853] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199161, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.167468] env[62368]: DEBUG nova.objects.instance [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'flavor' on Instance uuid fa637a2e-a113-4d23-8924-728ba75f202c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.173818] env[62368]: DEBUG nova.compute.manager [req-9301e41b-7931-49e0-b694-c9ad77de9c13 req-88a54586-aaa9-4361-9620-279e9f7f5766 service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Received event network-changed-11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.174085] env[62368]: DEBUG nova.compute.manager [req-9301e41b-7931-49e0-b694-c9ad77de9c13 req-88a54586-aaa9-4361-9620-279e9f7f5766 service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Refreshing instance network info cache due to event network-changed-11de7a9b-926a-47f0-a28d-a30419725a7f. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1080.174306] env[62368]: DEBUG oslo_concurrency.lockutils [req-9301e41b-7931-49e0-b694-c9ad77de9c13 req-88a54586-aaa9-4361-9620-279e9f7f5766 service nova] Acquiring lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.174459] env[62368]: DEBUG oslo_concurrency.lockutils [req-9301e41b-7931-49e0-b694-c9ad77de9c13 req-88a54586-aaa9-4361-9620-279e9f7f5766 service nova] Acquired lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.174661] env[62368]: DEBUG nova.network.neutron [req-9301e41b-7931-49e0-b694-c9ad77de9c13 req-88a54586-aaa9-4361-9620-279e9f7f5766 service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Refreshing network info cache for port 11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1080.268016] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199159, 'name': ReconfigVM_Task, 'duration_secs': 1.039} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.268333] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3/988cac3a-1637-4682-ba30-e9e1f73cc3e3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.268973] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c9ca508-0d85-48db-b54d-cc6bce12d2c2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.274966] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1080.274966] env[62368]: value = "task-1199162" [ 1080.274966] env[62368]: _type = "Task" [ 1080.274966] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.282954] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199162, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.530807] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 58157ab1-80a4-427c-812b-f6fde1f8db68] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.657392] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199161, 'name': CreateVM_Task, 'duration_secs': 0.307137} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.657571] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1080.658352] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'guest_format': None, 'attachment_id': '3852ace7-918a-4578-9cd0-b4c4ec4a9801', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259906', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'name': 'volume-a5a81572-db8f-4766-b46a-5e14fa56e64c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753bbfe2-405e-4e81-8756-9f52cb637969', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'serial': 'a5a81572-db8f-4766-b46a-5e14fa56e64c'}, 'mount_device': '/dev/sda', 'boot_index': 0, 'disk_bus': None, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=62368) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1080.658584] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Root volume attach. Driver type: vmdk {{(pid=62368) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1080.659385] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a893caf-3ace-441a-941e-ea08ec3f28dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.666714] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a010ac5e-7027-45a4-bce7-668d021ebfec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.673941] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd374646-60fa-4177-9711-f96e0366e820 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.681693] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-96c5a750-959f-4e1b-8722-c0bb404bf19c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.688785] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1080.688785] env[62368]: value = "task-1199163" [ 1080.688785] env[62368]: _type = "Task" [ 1080.688785] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.696383] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199163, 'name': RelocateVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.785320] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199162, 'name': Rename_Task, 'duration_secs': 0.131384} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.786082] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1080.786082] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e3ea9d9-080b-4338-ad1e-72d00520cfc0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.792016] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1080.792016] env[62368]: value = "task-1199164" [ 1080.792016] env[62368]: _type = "Task" [ 1080.792016] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.800658] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199164, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.926900] env[62368]: DEBUG nova.network.neutron [req-9301e41b-7931-49e0-b694-c9ad77de9c13 req-88a54586-aaa9-4361-9620-279e9f7f5766 service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updated VIF entry in instance network info cache for port 11de7a9b-926a-47f0-a28d-a30419725a7f. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1080.927342] env[62368]: DEBUG nova.network.neutron [req-9301e41b-7931-49e0-b694-c9ad77de9c13 req-88a54586-aaa9-4361-9620-279e9f7f5766 service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance_info_cache with network_info: [{"id": "11de7a9b-926a-47f0-a28d-a30419725a7f", "address": "fa:16:3e:b0:ab:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11de7a9b-92", "ovs_interfaceid": "11de7a9b-926a-47f0-a28d-a30419725a7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.033816] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: eea21546-fbbf-4440-829c-8583c4ccabb6] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.177462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-839041f1-4276-4271-a880-a17eaa03e370 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.230s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.198351] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199163, 'name': RelocateVM_Task, 'duration_secs': 0.359458} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.198625] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1081.198875] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259906', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'name': 'volume-a5a81572-db8f-4766-b46a-5e14fa56e64c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753bbfe2-405e-4e81-8756-9f52cb637969', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'serial': 'a5a81572-db8f-4766-b46a-5e14fa56e64c'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1081.199684] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2072cbfd-3e7a-4770-8e2d-1411938ef84b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.215839] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4d3960-ef55-4c87-9194-c2d090492a86 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.237539] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-a5a81572-db8f-4766-b46a-5e14fa56e64c/volume-a5a81572-db8f-4766-b46a-5e14fa56e64c.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.237787] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4021d679-c4ae-4bde-9c61-16005ba8bf69 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.256791] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1081.256791] env[62368]: value = "task-1199165" [ 1081.256791] env[62368]: _type = "Task" [ 1081.256791] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.264078] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.300984] env[62368]: DEBUG oslo_vmware.api [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199164, 'name': PowerOnVM_Task, 'duration_secs': 0.409641} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.301871] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1081.301871] env[62368]: DEBUG nova.compute.manager [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1081.302464] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e34f762-e0b2-4bdc-ae81-24a204dec80d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.430579] env[62368]: DEBUG oslo_concurrency.lockutils [req-9301e41b-7931-49e0-b694-c9ad77de9c13 req-88a54586-aaa9-4361-9620-279e9f7f5766 service nova] Releasing lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.537092] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 54083854-e314-44a7-b4b0-fbef5fa6b1ef] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.767670] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199165, 'name': ReconfigVM_Task, 'duration_secs': 0.235465} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.768056] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-a5a81572-db8f-4766-b46a-5e14fa56e64c/volume-a5a81572-db8f-4766-b46a-5e14fa56e64c.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.773337] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aaa7674e-c0e8-4044-9e6e-7d76f0e40aab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.788998] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1081.788998] env[62368]: value = "task-1199166" [ 1081.788998] env[62368]: _type = "Task" [ 1081.788998] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.799027] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199166, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.817342] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.817580] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.817765] env[62368]: DEBUG nova.objects.instance [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62368) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1082.040588] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b79f0e79-9e3a-47c7-9949-8743601ec6c5] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.094506] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.094767] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.095047] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.095257] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.095435] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.097415] env[62368]: INFO nova.compute.manager [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Terminating instance [ 1082.099175] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "refresh_cache-988cac3a-1637-4682-ba30-e9e1f73cc3e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.099344] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired lock "refresh_cache-988cac3a-1637-4682-ba30-e9e1f73cc3e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.099515] env[62368]: DEBUG nova.network.neutron [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1082.199573] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.199854] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.200067] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.200430] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.200645] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.202800] env[62368]: INFO nova.compute.manager [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Terminating instance [ 1082.204523] env[62368]: DEBUG nova.compute.manager [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1082.204733] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1082.205563] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5141cdc6-c5e5-4cdb-8f7b-9d53ac30e2a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.213422] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1082.213583] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d5fc79f-855b-4bd5-9453-2089727d1c49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.219728] env[62368]: DEBUG oslo_vmware.api [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1082.219728] env[62368]: value = "task-1199167" [ 1082.219728] env[62368]: _type = "Task" [ 1082.219728] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.227927] env[62368]: DEBUG oslo_vmware.api [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.298536] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199166, 'name': ReconfigVM_Task, 'duration_secs': 0.115441} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.298830] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259906', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'name': 'volume-a5a81572-db8f-4766-b46a-5e14fa56e64c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753bbfe2-405e-4e81-8756-9f52cb637969', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'serial': 'a5a81572-db8f-4766-b46a-5e14fa56e64c'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1082.299319] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd5d6324-90c7-4f23-ad73-a705f08f984b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.305691] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1082.305691] env[62368]: value = "task-1199168" [ 1082.305691] env[62368]: _type = "Task" [ 1082.305691] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.313572] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199168, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.543834] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a0cea538-b162-4504-ac34-803a2d5a8071] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.618604] env[62368]: DEBUG nova.network.neutron [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.663307] env[62368]: DEBUG nova.network.neutron [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.731270] env[62368]: DEBUG oslo_vmware.api [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199167, 'name': PowerOffVM_Task, 'duration_secs': 0.221472} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.731548] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1082.731721] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1082.731983] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef5655c8-0d64-4465-b78c-dd9170ae34a7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.794546] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1082.794834] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1082.795042] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleting the datastore file [datastore1] fa637a2e-a113-4d23-8924-728ba75f202c {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1082.795328] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09d57256-efaa-4d1d-aaf1-8cf2047a54eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.802013] env[62368]: DEBUG oslo_vmware.api [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1082.802013] env[62368]: value = "task-1199170" [ 1082.802013] env[62368]: _type = "Task" [ 1082.802013] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.811632] env[62368]: DEBUG oslo_vmware.api [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.816382] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199168, 'name': Rename_Task, 'duration_secs': 0.143487} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.816627] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1082.816887] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b1f88dc-7d3f-4501-b8ed-c45f499d1a42 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.821847] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1082.821847] env[62368]: value = "task-1199171" [ 1082.821847] env[62368]: _type = "Task" [ 1082.821847] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.826949] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e13db909-ea80-4984-9396-8395ae41007e tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.830535] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199171, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.047063] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 87dc392f-7bba-44bb-ba38-2fd4e1c6129e] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.165623] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Releasing lock "refresh_cache-988cac3a-1637-4682-ba30-e9e1f73cc3e3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.166078] env[62368]: DEBUG nova.compute.manager [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1083.166288] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1083.167245] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021d7b2d-c7da-4a27-92d1-5a4b59a6b1f8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.176369] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1083.176616] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b216cee-7234-4344-8a08-4cd6ebc0b3ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.183707] env[62368]: DEBUG oslo_vmware.api [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1083.183707] env[62368]: value = "task-1199172" [ 1083.183707] env[62368]: _type = "Task" [ 1083.183707] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.191965] env[62368]: DEBUG oslo_vmware.api [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199172, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.312529] env[62368]: DEBUG oslo_vmware.api [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12474} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.312818] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1083.313030] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1083.313214] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1083.313393] env[62368]: INFO nova.compute.manager [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1083.313643] env[62368]: DEBUG oslo.service.loopingcall [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.313846] env[62368]: DEBUG nova.compute.manager [-] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1083.313940] env[62368]: DEBUG nova.network.neutron [-] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1083.331734] env[62368]: DEBUG oslo_vmware.api [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199171, 'name': PowerOnVM_Task, 'duration_secs': 0.418029} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.331998] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1083.332224] env[62368]: INFO nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Took 3.66 seconds to spawn the instance on the hypervisor. [ 1083.332409] env[62368]: DEBUG nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1083.333209] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c260419-4f6f-4bb3-a566-42cb699231ab {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.551055] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 8aa55880-3dfe-4c1b-9aa2-3c4a5926a77d] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.695115] env[62368]: DEBUG oslo_vmware.api [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199172, 'name': PowerOffVM_Task, 'duration_secs': 0.124289} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.695396] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1083.695567] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1083.695821] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0501c3df-9fd9-459b-8f77-60f7e3a154f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.721638] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1083.721893] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1083.722060] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Deleting the datastore file [datastore2] 988cac3a-1637-4682-ba30-e9e1f73cc3e3 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1083.722397] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f3356fe-6844-44bb-8051-c90ab6a677d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.728264] env[62368]: DEBUG nova.compute.manager [req-754d2829-4934-4857-86da-c0e2bebbeb21 req-5eb19728-8a44-4556-8bbb-dafa98615255 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Received event network-vif-deleted-15eed45d-e09b-4aac-b3e2-e3b77ed1efe2 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.729631] env[62368]: INFO nova.compute.manager [req-754d2829-4934-4857-86da-c0e2bebbeb21 req-5eb19728-8a44-4556-8bbb-dafa98615255 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Neutron deleted interface 15eed45d-e09b-4aac-b3e2-e3b77ed1efe2; detaching it from the instance and deleting it from the info cache [ 1083.729631] env[62368]: DEBUG nova.network.neutron [req-754d2829-4934-4857-86da-c0e2bebbeb21 req-5eb19728-8a44-4556-8bbb-dafa98615255 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.731836] env[62368]: DEBUG oslo_vmware.api [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1083.731836] env[62368]: value = "task-1199174" [ 1083.731836] env[62368]: _type = "Task" [ 1083.731836] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.741335] env[62368]: DEBUG oslo_vmware.api [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.851995] env[62368]: INFO nova.compute.manager [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Took 10.13 seconds to build instance. [ 1084.054405] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 0f1bff84-1fb9-471a-b685-7d527b376b0a] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.204385] env[62368]: DEBUG nova.network.neutron [-] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.232872] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9609aa5-92f4-4348-8720-a612dd00ad90 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.246546] env[62368]: DEBUG oslo_vmware.api [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132587} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.248032] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1084.248271] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1084.248461] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1084.248640] env[62368]: INFO nova.compute.manager [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1084.248901] env[62368]: DEBUG oslo.service.loopingcall [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.249580] env[62368]: DEBUG nova.compute.manager [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1084.249689] env[62368]: DEBUG nova.network.neutron [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1084.254029] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563709fe-a3ae-4361-9d59-418a67435822 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.286677] env[62368]: DEBUG nova.compute.manager [req-754d2829-4934-4857-86da-c0e2bebbeb21 req-5eb19728-8a44-4556-8bbb-dafa98615255 service nova] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Detach interface failed, port_id=15eed45d-e09b-4aac-b3e2-e3b77ed1efe2, reason: Instance fa637a2e-a113-4d23-8924-728ba75f202c could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1084.287654] env[62368]: DEBUG nova.network.neutron [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1084.354129] env[62368]: DEBUG oslo_concurrency.lockutils [None req-1fd805fb-36ba-4a1f-9e1e-3dc688789c81 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.638s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.559022] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 67d8c43a-d3af-407b-847a-1c2c8c53dbfb] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.707553] env[62368]: INFO nova.compute.manager [-] [instance: fa637a2e-a113-4d23-8924-728ba75f202c] Took 1.39 seconds to deallocate network for instance. [ 1084.791776] env[62368]: DEBUG nova.network.neutron [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.972509] env[62368]: DEBUG nova.compute.manager [req-046be526-0ad7-4e5a-98d7-d22ee5722f74 req-c1650692-98c2-4969-b71d-7ee21b4ce580 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Received event network-changed-b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1084.972715] env[62368]: DEBUG nova.compute.manager [req-046be526-0ad7-4e5a-98d7-d22ee5722f74 req-c1650692-98c2-4969-b71d-7ee21b4ce580 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Refreshing instance network info cache due to event network-changed-b4fab9b2-299a-4e59-92bf-efdc3d583416. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1084.972941] env[62368]: DEBUG oslo_concurrency.lockutils [req-046be526-0ad7-4e5a-98d7-d22ee5722f74 req-c1650692-98c2-4969-b71d-7ee21b4ce580 service nova] Acquiring lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.973764] env[62368]: DEBUG oslo_concurrency.lockutils [req-046be526-0ad7-4e5a-98d7-d22ee5722f74 req-c1650692-98c2-4969-b71d-7ee21b4ce580 service nova] Acquired lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.973860] env[62368]: DEBUG nova.network.neutron [req-046be526-0ad7-4e5a-98d7-d22ee5722f74 req-c1650692-98c2-4969-b71d-7ee21b4ce580 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Refreshing network info cache for port b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1085.064831] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 3f1af54b-392f-432a-9ffa-a133da428f94] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.216047] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.216367] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.216544] env[62368]: DEBUG nova.objects.instance [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'resources' on Instance uuid fa637a2e-a113-4d23-8924-728ba75f202c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.295093] env[62368]: INFO nova.compute.manager [-] [instance: 988cac3a-1637-4682-ba30-e9e1f73cc3e3] Took 1.04 seconds to deallocate network for instance. [ 1085.565632] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 300c6c32-b365-40c2-bc7d-6ce028f10476] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.675644] env[62368]: DEBUG nova.network.neutron [req-046be526-0ad7-4e5a-98d7-d22ee5722f74 req-c1650692-98c2-4969-b71d-7ee21b4ce580 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updated VIF entry in instance network info cache for port b4fab9b2-299a-4e59-92bf-efdc3d583416. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1085.676332] env[62368]: DEBUG nova.network.neutron [req-046be526-0ad7-4e5a-98d7-d22ee5722f74 req-c1650692-98c2-4969-b71d-7ee21b4ce580 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updating instance_info_cache with network_info: [{"id": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "address": "fa:16:3e:2d:06:3b", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4fab9b2-29", "ovs_interfaceid": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.801272] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.830657] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f642dc74-f38b-4ffb-a2f7-c28699aff1e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.838533] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1fe3c6-53c3-4a1f-a592-df165660f437 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.870916] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9905cb-745d-49f8-a012-e512d59e1105 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.882650] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4341a738-67f6-4a65-839c-a789a7f221e1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.898244] env[62368]: DEBUG nova.compute.provider_tree [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.068883] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 0ef4b1bc-9f25-4a7d-8acd-55181a7d57e3] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.179397] env[62368]: DEBUG oslo_concurrency.lockutils [req-046be526-0ad7-4e5a-98d7-d22ee5722f74 req-c1650692-98c2-4969-b71d-7ee21b4ce580 service nova] Releasing lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.382736] env[62368]: DEBUG nova.compute.manager [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Stashing vm_state: active {{(pid=62368) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1086.401050] env[62368]: DEBUG nova.scheduler.client.report [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.573872] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: af36f8e3-dffb-4cfe-a8fb-3ecad36e0a03] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.901739] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.907030] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.689s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.907709] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.107s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.907961] env[62368]: DEBUG nova.objects.instance [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lazy-loading 'resources' on Instance uuid 988cac3a-1637-4682-ba30-e9e1f73cc3e3 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.922909] env[62368]: INFO nova.scheduler.client.report [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted allocations for instance fa637a2e-a113-4d23-8924-728ba75f202c [ 1086.996626] env[62368]: DEBUG nova.compute.manager [req-41176d23-4d65-43ee-8da7-5aa718c35825 req-aa054743-36f8-4158-8e2c-bfe5418669ee service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Received event network-changed-11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1086.996626] env[62368]: DEBUG nova.compute.manager [req-41176d23-4d65-43ee-8da7-5aa718c35825 req-aa054743-36f8-4158-8e2c-bfe5418669ee service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Refreshing instance network info cache due to event network-changed-11de7a9b-926a-47f0-a28d-a30419725a7f. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1086.996907] env[62368]: DEBUG oslo_concurrency.lockutils [req-41176d23-4d65-43ee-8da7-5aa718c35825 req-aa054743-36f8-4158-8e2c-bfe5418669ee service nova] Acquiring lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.997078] env[62368]: DEBUG oslo_concurrency.lockutils [req-41176d23-4d65-43ee-8da7-5aa718c35825 req-aa054743-36f8-4158-8e2c-bfe5418669ee service nova] Acquired lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.997255] env[62368]: DEBUG nova.network.neutron [req-41176d23-4d65-43ee-8da7-5aa718c35825 req-aa054743-36f8-4158-8e2c-bfe5418669ee service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Refreshing network info cache for port 11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1087.076184] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 417f000f-cf23-404d-877c-45990d1a7c77] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.434322] env[62368]: DEBUG oslo_concurrency.lockutils [None req-bef87c61-9bb0-4450-89b5-9b0c6c0c382e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "fa637a2e-a113-4d23-8924-728ba75f202c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.234s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.508890] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b09b300-3a35-4a80-b766-ddb126419a10 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.517332] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356083f2-a02d-4e71-a387-e6d3515d9e0b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.548841] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb3c4b1-81bb-415c-b204-2307c2c8f2da {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.555813] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0997fc91-ee42-4726-80fc-2ee3d6087aed {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.568543] env[62368]: DEBUG nova.compute.provider_tree [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.579555] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 4fe7b1a2-f894-4131-9456-fb4df8a4532d] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.782404] env[62368]: DEBUG nova.network.neutron [req-41176d23-4d65-43ee-8da7-5aa718c35825 req-aa054743-36f8-4158-8e2c-bfe5418669ee service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updated VIF entry in instance network info cache for port 11de7a9b-926a-47f0-a28d-a30419725a7f. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1087.782404] env[62368]: DEBUG nova.network.neutron [req-41176d23-4d65-43ee-8da7-5aa718c35825 req-aa054743-36f8-4158-8e2c-bfe5418669ee service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance_info_cache with network_info: [{"id": "11de7a9b-926a-47f0-a28d-a30419725a7f", "address": "fa:16:3e:b0:ab:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11de7a9b-92", "ovs_interfaceid": "11de7a9b-926a-47f0-a28d-a30419725a7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.071933] env[62368]: DEBUG nova.scheduler.client.report [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.082814] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 13765305-2e55-4ee8-9a6f-4ae5ee724367] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.285383] env[62368]: DEBUG oslo_concurrency.lockutils [req-41176d23-4d65-43ee-8da7-5aa718c35825 req-aa054743-36f8-4158-8e2c-bfe5418669ee service nova] Releasing lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.579013] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.581382] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.680s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.585381] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 085ebe93-aa24-4626-94fe-241c4297e4db] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.606782] env[62368]: INFO nova.scheduler.client.report [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Deleted allocations for instance 988cac3a-1637-4682-ba30-e9e1f73cc3e3 [ 1089.087206] env[62368]: INFO nova.compute.claims [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1089.092156] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b393f654-0611-493e-a89f-445dfd3f8c13] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.113963] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0531120-f62e-466b-9f12-876a2280888d tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "988cac3a-1637-4682-ba30-e9e1f73cc3e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.019s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.567214] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.567469] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.596127] env[62368]: INFO nova.compute.resource_tracker [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating resource usage from migration b4377ef5-e1ad-408d-99f2-77d5a4fcf6cc [ 1089.598777] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 9ce5c5a1-5c90-4f76-a511-d8f855b6bab4] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.645796] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.646055] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.646282] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.646472] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.646643] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.651301] env[62368]: INFO nova.compute.manager [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Terminating instance [ 1089.653396] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "refresh_cache-eb9281af-c914-4cf1-bc32-4cf43fb98c81" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.653396] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquired lock "refresh_cache-eb9281af-c914-4cf1-bc32-4cf43fb98c81" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.653396] env[62368]: DEBUG nova.network.neutron [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1089.699220] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf24e5d5-2f34-4f62-8a1f-ca68aa8d039d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.707039] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cdbe34-95e8-4515-8d14-29f6fd127241 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.737840] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5081d0b0-d03e-4732-a327-d4327b1ed4aa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.744760] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731a10e1-cbcf-47d9-a350-8c50ec442ada {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.757545] env[62368]: DEBUG nova.compute.provider_tree [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.070109] env[62368]: DEBUG nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1090.101627] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 4dc7b6b4-1bf5-4195-bb93-14756f8f9986] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.171834] env[62368]: DEBUG nova.network.neutron [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1090.217664] env[62368]: DEBUG nova.network.neutron [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.261111] env[62368]: DEBUG nova.scheduler.client.report [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1090.593466] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.605710] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: e4a0d368-6afc-4dd1-9c6b-8d8cfeea4ce0] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.721024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Releasing lock "refresh_cache-eb9281af-c914-4cf1-bc32-4cf43fb98c81" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.721622] env[62368]: DEBUG nova.compute.manager [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1090.721837] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1090.722757] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f442df0-85f5-4d6a-951b-e91dd885e7a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.731288] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1090.731541] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7cb24ac-ca7a-42cd-b255-202bc24c2c88 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.738744] env[62368]: DEBUG oslo_vmware.api [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1090.738744] env[62368]: value = "task-1199175" [ 1090.738744] env[62368]: _type = "Task" [ 1090.738744] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.745770] env[62368]: DEBUG oslo_vmware.api [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199175, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.766386] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.185s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.766609] env[62368]: INFO nova.compute.manager [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Migrating [ 1090.773505] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.180s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.775431] env[62368]: INFO nova.compute.claims [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1091.112398] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 3f36000e-b93d-4dda-ac39-b8459203c227] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.249408] env[62368]: DEBUG oslo_vmware.api [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199175, 'name': PowerOffVM_Task, 'duration_secs': 0.127819} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.249691] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1091.249900] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1091.250183] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23d42346-83c1-4db7-9fb2-7b32ec8a3a40 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.275390] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1091.275500] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1091.275655] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Deleting the datastore file [datastore2] eb9281af-c914-4cf1-bc32-4cf43fb98c81 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.275910] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45996eb6-decd-4954-82df-d7ed42ebc682 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.282239] env[62368]: DEBUG oslo_vmware.api [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for the task: (returnval){ [ 1091.282239] env[62368]: value = "task-1199177" [ 1091.282239] env[62368]: _type = "Task" [ 1091.282239] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.290672] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.290843] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquired lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.291235] env[62368]: DEBUG nova.network.neutron [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1091.292886] env[62368]: DEBUG oslo_vmware.api [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199177, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.615337] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 8b7ad7ed-8f2d-4c76-b335-4a2ce5327072] Instance has had 0 of 5 cleanup attempts {{(pid=62368) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.689230] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b4070763-a2f2-480f-909c-e323f15a9ac5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.689798] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.791909] env[62368]: DEBUG oslo_vmware.api [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Task: {'id': task-1199177, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1829} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.792210] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.792418] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1091.792606] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1091.792785] env[62368]: INFO nova.compute.manager [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1091.793041] env[62368]: DEBUG oslo.service.loopingcall [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.795785] env[62368]: DEBUG nova.compute.manager [-] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.795880] env[62368]: DEBUG nova.network.neutron [-] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1091.813505] env[62368]: DEBUG nova.network.neutron [-] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1091.900480] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98a06ed-0759-46aa-b12e-7159155d25e1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.910251] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443d0804-e5de-4908-9cb9-3fe430e507f2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.943409] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2079c187-05d4-4c90-8ff2-2c8bd98e669d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.951913] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68579101-751c-44a2-b303-2902b585d466 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.965878] env[62368]: DEBUG nova.compute.provider_tree [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.070326] env[62368]: DEBUG nova.network.neutron [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance_info_cache with network_info: [{"id": "11de7a9b-926a-47f0-a28d-a30419725a7f", "address": "fa:16:3e:b0:ab:4e", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11de7a9b-92", "ovs_interfaceid": "11de7a9b-926a-47f0-a28d-a30419725a7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.119045] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.119352] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Cleaning up deleted instances with incomplete migration {{(pid=62368) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1092.194037] env[62368]: DEBUG nova.compute.utils [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1092.317263] env[62368]: DEBUG nova.network.neutron [-] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.468495] env[62368]: DEBUG nova.scheduler.client.report [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.573216] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Releasing lock "refresh_cache-753bbfe2-405e-4e81-8756-9f52cb637969" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.696291] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.819451] env[62368]: INFO nova.compute.manager [-] [instance: eb9281af-c914-4cf1-bc32-4cf43fb98c81] Took 1.02 seconds to deallocate network for instance. [ 1092.973270] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.200s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.973802] env[62368]: DEBUG nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1093.326120] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.326443] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.326628] env[62368]: DEBUG nova.objects.instance [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lazy-loading 'resources' on Instance uuid eb9281af-c914-4cf1-bc32-4cf43fb98c81 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.479778] env[62368]: DEBUG nova.compute.utils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1093.481160] env[62368]: DEBUG nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1093.481345] env[62368]: DEBUG nova.network.neutron [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1093.527872] env[62368]: DEBUG nova.policy [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9c6e1b84100c4d4bb47f5d95272b62b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b7091fbcb2f4879b78075ed06b95c0d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1093.766640] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b4070763-a2f2-480f-909c-e323f15a9ac5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.766993] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.767276] env[62368]: INFO nova.compute.manager [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Attaching volume cb82f49c-702d-4495-9424-daff96816034 to /dev/sdb [ 1093.788052] env[62368]: DEBUG nova.network.neutron [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Successfully created port: 7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1093.809940] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb46c269-918a-4a05-92b4-0f3dfc283cd1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.815767] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18278015-77bb-4d3f-ada1-2be251856598 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.830972] env[62368]: DEBUG nova.virt.block_device [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updating existing volume attachment record: 6e52a94e-d15e-41a6-83d6-a6fb95a7728e {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1093.941963] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42b5b7c-ce49-4c5c-b460-4419e6f74dad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.949467] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004f9daa-e509-4627-9a2a-fc6b9372f9fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.980124] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd165236-f0cc-487f-ab1d-a3ea3f2c4812 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.984338] env[62368]: DEBUG nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1093.990528] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a92db1b-cf3e-4a6e-9424-3748829fd93f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.004007] env[62368]: DEBUG nova.compute.provider_tree [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.089905] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2b0992-2205-4fef-80e9-60d6b868cf8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.108314] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance '753bbfe2-405e-4e81-8756-9f52cb637969' progress to 0 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1094.119322] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.119540] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.507033] env[62368]: DEBUG nova.scheduler.client.report [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1094.614069] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1094.614383] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0fef2883-3e56-4c8d-a302-d285a6761b48 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.624988] env[62368]: DEBUG oslo_vmware.api [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1094.624988] env[62368]: value = "task-1199181" [ 1094.624988] env[62368]: _type = "Task" [ 1094.624988] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.625474] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.625631] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1094.625751] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Rebuilding the list of instances to heal {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1094.635879] env[62368]: DEBUG oslo_vmware.api [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199181, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.997525] env[62368]: DEBUG nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1095.011992] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.685s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.024249] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1095.024416] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1095.024581] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1095.024767] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1095.024944] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1095.025394] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1095.025394] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1095.025513] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1095.025686] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1095.025877] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1095.026074] env[62368]: DEBUG nova.virt.hardware [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1095.026957] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4d46c2-935b-439f-b493-19d64d61e537 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.032627] env[62368]: INFO nova.scheduler.client.report [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Deleted allocations for instance eb9281af-c914-4cf1-bc32-4cf43fb98c81 [ 1095.038127] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa16b68-9584-453c-9093-b5d1056b8526 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.133575] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1095.138629] env[62368]: DEBUG oslo_vmware.api [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199181, 'name': PowerOffVM_Task, 'duration_secs': 0.180927} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.139184] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1095.139408] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance '753bbfe2-405e-4e81-8756-9f52cb637969' progress to 17 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1095.164913] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.165077] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquired lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.165529] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Forcefully refreshing network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1095.165529] env[62368]: DEBUG nova.objects.instance [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lazy-loading 'info_cache' on Instance uuid 2a46c954-449e-4d62-be80-add1040ed4c0 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.170226] env[62368]: DEBUG nova.compute.manager [req-025c2740-aa85-4265-8b10-109f7eb8ed24 req-af64736c-760b-4318-9698-b4a6debf1c70 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received event network-vif-plugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.170434] env[62368]: DEBUG oslo_concurrency.lockutils [req-025c2740-aa85-4265-8b10-109f7eb8ed24 req-af64736c-760b-4318-9698-b4a6debf1c70 service nova] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.170639] env[62368]: DEBUG oslo_concurrency.lockutils [req-025c2740-aa85-4265-8b10-109f7eb8ed24 req-af64736c-760b-4318-9698-b4a6debf1c70 service nova] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.170805] env[62368]: DEBUG oslo_concurrency.lockutils [req-025c2740-aa85-4265-8b10-109f7eb8ed24 req-af64736c-760b-4318-9698-b4a6debf1c70 service nova] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.170977] env[62368]: DEBUG nova.compute.manager [req-025c2740-aa85-4265-8b10-109f7eb8ed24 req-af64736c-760b-4318-9698-b4a6debf1c70 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] No waiting events found dispatching network-vif-plugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1095.171203] env[62368]: WARNING nova.compute.manager [req-025c2740-aa85-4265-8b10-109f7eb8ed24 req-af64736c-760b-4318-9698-b4a6debf1c70 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received unexpected event network-vif-plugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 for instance with vm_state building and task_state spawning. [ 1095.257902] env[62368]: DEBUG nova.network.neutron [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Successfully updated port: 7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1095.545929] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c4e27972-3536-4001-8b89-bbf9629b9dba tempest-ServerShowV247Test-485517136 tempest-ServerShowV247Test-485517136-project-member] Lock "eb9281af-c914-4cf1-bc32-4cf43fb98c81" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.900s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.645961] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1095.645961] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1095.645961] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1095.645961] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1095.646298] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1095.646298] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1095.646404] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1095.646842] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1095.646842] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1095.646941] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1095.647116] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1095.652134] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c10877e-f87e-4dfe-9be4-fe5e8e9b6737 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.669726] env[62368]: DEBUG oslo_vmware.api [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1095.669726] env[62368]: value = "task-1199182" [ 1095.669726] env[62368]: _type = "Task" [ 1095.669726] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.677699] env[62368]: DEBUG oslo_vmware.api [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199182, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.760538] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.760696] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.760867] env[62368]: DEBUG nova.network.neutron [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1096.181767] env[62368]: DEBUG oslo_vmware.api [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199182, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.293441] env[62368]: DEBUG nova.network.neutron [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1096.431728] env[62368]: DEBUG nova.network.neutron [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [{"id": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "address": "fa:16:3e:0c:7d:cb", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8eb12d-c6", "ovs_interfaceid": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.682395] env[62368]: DEBUG oslo_vmware.api [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199182, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.865187] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updating instance_info_cache with network_info: [{"id": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "address": "fa:16:3e:2d:06:3b", "network": {"id": "b58ee7d0-4fe3-40cc-99de-f0d864dfa5a6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1554806746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "430e72f3b08a4207a2fb4465ff160714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4fab9b2-29", "ovs_interfaceid": "b4fab9b2-299a-4e59-92bf-efdc3d583416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.933893] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.934224] env[62368]: DEBUG nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Instance network_info: |[{"id": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "address": "fa:16:3e:0c:7d:cb", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8eb12d-c6", "ovs_interfaceid": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1096.934920] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:7d:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e8eb12d-c67a-4b59-9e66-b3b530e66e75', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1096.943210] env[62368]: DEBUG oslo.service.loopingcall [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1096.943721] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1096.943967] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28ef9a18-e288-4e78-9d33-128b79685e5b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.964735] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1096.964735] env[62368]: value = "task-1199184" [ 1096.964735] env[62368]: _type = "Task" [ 1096.964735] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.972611] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199184, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.183954] env[62368]: DEBUG oslo_vmware.api [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199182, 'name': ReconfigVM_Task, 'duration_secs': 1.172413} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.184287] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance '753bbfe2-405e-4e81-8756-9f52cb637969' progress to 33 {{(pid=62368) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1097.195743] env[62368]: DEBUG nova.compute.manager [req-a5340543-93a9-42d3-af6d-b9c2771421d0 req-8587f579-b05c-4802-a239-7fe87e335b0d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received event network-changed-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1097.195941] env[62368]: DEBUG nova.compute.manager [req-a5340543-93a9-42d3-af6d-b9c2771421d0 req-8587f579-b05c-4802-a239-7fe87e335b0d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Refreshing instance network info cache due to event network-changed-7e8eb12d-c67a-4b59-9e66-b3b530e66e75. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1097.196172] env[62368]: DEBUG oslo_concurrency.lockutils [req-a5340543-93a9-42d3-af6d-b9c2771421d0 req-8587f579-b05c-4802-a239-7fe87e335b0d service nova] Acquiring lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.196323] env[62368]: DEBUG oslo_concurrency.lockutils [req-a5340543-93a9-42d3-af6d-b9c2771421d0 req-8587f579-b05c-4802-a239-7fe87e335b0d service nova] Acquired lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.196526] env[62368]: DEBUG nova.network.neutron [req-a5340543-93a9-42d3-af6d-b9c2771421d0 req-8587f579-b05c-4802-a239-7fe87e335b0d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Refreshing network info cache for port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1097.367933] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Releasing lock "refresh_cache-2a46c954-449e-4d62-be80-add1040ed4c0" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.368205] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updated the network info_cache for instance {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1097.368412] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.368575] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.368724] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.368876] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.369030] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.369185] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.369339] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1097.369485] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.473874] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199184, 'name': CreateVM_Task, 'duration_secs': 0.293534} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.474010] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1097.474694] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.474864] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.475228] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1097.475474] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35d8ba3a-9931-46e9-a4f5-ab1625dc1189 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.479487] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1097.479487] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52bf07dc-fb44-03ee-9e84-5e082f5a6dd4" [ 1097.479487] env[62368]: _type = "Task" [ 1097.479487] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.487669] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52bf07dc-fb44-03ee-9e84-5e082f5a6dd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.691033] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1097.691383] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1097.691383] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1097.691549] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1097.691701] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1097.692462] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1097.692462] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1097.692462] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1097.692462] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1097.692656] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1097.692775] env[62368]: DEBUG nova.virt.hardware [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1097.873622] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.873874] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.874093] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.874281] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1097.875409] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaee8bfa-a082-4ee4-bffd-e36b09e5a111 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.883808] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55ab16a-0a6e-414d-a383-23fe8de1aaac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.897200] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e63447e-9824-4748-b05a-eafdf9339c8f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.903371] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338d21e7-fe42-412e-88c4-001ee1c454e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.907965] env[62368]: DEBUG nova.network.neutron [req-a5340543-93a9-42d3-af6d-b9c2771421d0 req-8587f579-b05c-4802-a239-7fe87e335b0d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updated VIF entry in instance network info cache for port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1097.908336] env[62368]: DEBUG nova.network.neutron [req-a5340543-93a9-42d3-af6d-b9c2771421d0 req-8587f579-b05c-4802-a239-7fe87e335b0d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [{"id": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "address": "fa:16:3e:0c:7d:cb", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8eb12d-c6", "ovs_interfaceid": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.932340] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180242MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1097.932485] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.932664] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.989608] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52bf07dc-fb44-03ee-9e84-5e082f5a6dd4, 'name': SearchDatastore_Task, 'duration_secs': 0.008943} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.989894] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.990143] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.990382] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.990532] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.990711] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.990950] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65ac8441-0cea-4a8d-864b-1e61926c169b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.998980] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.999143] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1097.999812] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-907b3874-88da-4091-a001-d83ce90fc39c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.004379] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1098.004379] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5242cec1-962d-b798-cfbf-eca98e3f7c1e" [ 1098.004379] env[62368]: _type = "Task" [ 1098.004379] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.011190] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5242cec1-962d-b798-cfbf-eca98e3f7c1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.203135] env[62368]: ERROR nova.compute.manager [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Traceback (most recent call last): [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] yield [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] disk_info = self.driver.migrate_disk_and_power_off( [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] disk_key = device.key [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] AttributeError: 'NoneType' object has no attribute 'key' [ 1098.203135] env[62368]: ERROR nova.compute.manager [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] [ 1098.378048] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1098.378342] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259911', 'volume_id': 'cb82f49c-702d-4495-9424-daff96816034', 'name': 'volume-cb82f49c-702d-4495-9424-daff96816034', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b4070763-a2f2-480f-909c-e323f15a9ac5', 'attached_at': '', 'detached_at': '', 'volume_id': 'cb82f49c-702d-4495-9424-daff96816034', 'serial': 'cb82f49c-702d-4495-9424-daff96816034'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1098.379465] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1ed7a1-3114-4314-87bf-02256fa0cb78 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.395816] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa61d63-1eea-4ba1-beaa-5c8c6b050796 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.419229] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-cb82f49c-702d-4495-9424-daff96816034/volume-cb82f49c-702d-4495-9424-daff96816034.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.419471] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4d3d8b3-99e5-4ed0-9dee-780c5975ea47 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.435359] env[62368]: DEBUG oslo_concurrency.lockutils [req-a5340543-93a9-42d3-af6d-b9c2771421d0 req-8587f579-b05c-4802-a239-7fe87e335b0d service nova] Releasing lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.440210] env[62368]: DEBUG oslo_vmware.api [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1098.440210] env[62368]: value = "task-1199185" [ 1098.440210] env[62368]: _type = "Task" [ 1098.440210] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.447949] env[62368]: DEBUG oslo_vmware.api [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199185, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.514447] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5242cec1-962d-b798-cfbf-eca98e3f7c1e, 'name': SearchDatastore_Task, 'duration_secs': 0.008241} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.515524] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-677d23ed-4b16-4f55-909e-2f3f8162a1e3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.520163] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1098.520163] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529b2be1-062b-2348-403a-f7b15eadfcb7" [ 1098.520163] env[62368]: _type = "Task" [ 1098.520163] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.527376] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529b2be1-062b-2348-403a-f7b15eadfcb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.726811] env[62368]: INFO nova.compute.manager [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Swapping old allocation on dict_keys(['2202a74c-753d-4e1d-a031-7cefe24ee9d6']) held by migration b4377ef5-e1ad-408d-99f2-77d5a4fcf6cc for instance [ 1098.748428] env[62368]: DEBUG nova.scheduler.client.report [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Overwriting current allocation {'allocations': {'2202a74c-753d-4e1d-a031-7cefe24ee9d6': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 149}}, 'project_id': '430e72f3b08a4207a2fb4465ff160714', 'user_id': '87c366f937e2485c8b0dc317d6acb8f2', 'consumer_generation': 1} on consumer 753bbfe2-405e-4e81-8756-9f52cb637969 {{(pid=62368) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1098.941789] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Applying migration context for instance 753bbfe2-405e-4e81-8756-9f52cb637969 as it has an incoming, in-progress migration b4377ef5-e1ad-408d-99f2-77d5a4fcf6cc. Migration status is error {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1098.942777] env[62368]: INFO nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating resource usage from migration b4377ef5-e1ad-408d-99f2-77d5a4fcf6cc [ 1098.953856] env[62368]: DEBUG oslo_vmware.api [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199185, 'name': ReconfigVM_Task, 'duration_secs': 0.325897} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.954155] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-cb82f49c-702d-4495-9424-daff96816034/volume-cb82f49c-702d-4495-9424-daff96816034.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1098.959480] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 2a46c954-449e-4d62-be80-add1040ed4c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.959619] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance f5a33613-790b-4324-9b88-1c39f7c4269c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.959755] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b4070763-a2f2-480f-909c-e323f15a9ac5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.959881] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.960014] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 753bbfe2-405e-4e81-8756-9f52cb637969 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.960134] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.960317] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1098.960453] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1098.962466] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f6e63c7-acce-46e6-8aad-57968b3a890a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.977354] env[62368]: DEBUG oslo_vmware.api [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1098.977354] env[62368]: value = "task-1199186" [ 1098.977354] env[62368]: _type = "Task" [ 1098.977354] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.984834] env[62368]: DEBUG oslo_vmware.api [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199186, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.028719] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]529b2be1-062b-2348-403a-f7b15eadfcb7, 'name': SearchDatastore_Task, 'duration_secs': 0.00829} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.031095] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.031375] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1/27ff9bed-7c5f-4553-9d8e-8ec78708caf1.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1099.031819] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-113aeda9-c1ad-45e9-a141-20983cacb9ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.037621] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1099.037621] env[62368]: value = "task-1199187" [ 1099.037621] env[62368]: _type = "Task" [ 1099.037621] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.048117] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.056354] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db330316-8bc0-4c9d-bfcf-889e909a268b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.065571] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918931ab-b7c1-4eaf-92ab-b3aef36d08ff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.096277] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fdeb97-7653-4452-9a69-7aa0fb42542e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.103623] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb530b21-972d-4945-8390-dca827f7a4bf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.118242] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.488274] env[62368]: DEBUG oslo_vmware.api [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199186, 'name': ReconfigVM_Task, 'duration_secs': 0.149645} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.488590] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259911', 'volume_id': 'cb82f49c-702d-4495-9424-daff96816034', 'name': 'volume-cb82f49c-702d-4495-9424-daff96816034', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b4070763-a2f2-480f-909c-e323f15a9ac5', 'attached_at': '', 'detached_at': '', 'volume_id': 'cb82f49c-702d-4495-9424-daff96816034', 'serial': 'cb82f49c-702d-4495-9424-daff96816034'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1099.546920] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199187, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447837} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.547268] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1/27ff9bed-7c5f-4553-9d8e-8ec78708caf1.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1099.547502] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1099.547753] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63b85532-ed61-42df-a6ff-34db54c65ffb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.555182] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1099.555182] env[62368]: value = "task-1199188" [ 1099.555182] env[62368]: _type = "Task" [ 1099.555182] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.562788] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "753bbfe2-405e-4e81-8756-9f52cb637969" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.563069] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.563313] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.563512] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.563685] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.565251] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199188, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.565816] env[62368]: INFO nova.compute.manager [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Terminating instance [ 1099.568117] env[62368]: DEBUG nova.compute.manager [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1099.568316] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1099.568814] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c4150b8-9bd5-4908-9800-3b525e5a486e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.575244] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1099.575244] env[62368]: value = "task-1199189" [ 1099.575244] env[62368]: _type = "Task" [ 1099.575244] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.585323] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1099.585558] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1099.585784] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259906', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'name': 'volume-a5a81572-db8f-4766-b46a-5e14fa56e64c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753bbfe2-405e-4e81-8756-9f52cb637969', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'serial': 'a5a81572-db8f-4766-b46a-5e14fa56e64c'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1099.586532] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84ebb70-ab5b-4b62-95fa-639b71661c39 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.604711] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890d9474-f2fa-4c07-b0ea-79fd48dcd417 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.611035] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67f7574-93ff-47a6-9ff9-04f256544034 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.628256] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.631628] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89667a6d-56cf-4c6a-a4b6-b335b64094d7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.645904] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] The volume has not been displaced from its original location: [datastore2] volume-a5a81572-db8f-4766-b46a-5e14fa56e64c/volume-a5a81572-db8f-4766-b46a-5e14fa56e64c.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1099.651013] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1099.651774] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d50673d7-2409-4e59-b946-085e3c69edd9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.668544] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1099.668544] env[62368]: value = "task-1199190" [ 1099.668544] env[62368]: _type = "Task" [ 1099.668544] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.675808] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199190, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.065195] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199188, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064095} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.065519] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1100.066301] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11e9290-9be9-4a40-94c0-371f790a3b64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.090349] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1/27ff9bed-7c5f-4553-9d8e-8ec78708caf1.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1100.090615] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0441760e-4ee2-45d7-a5ac-32a51abc7b3a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.111239] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1100.111239] env[62368]: value = "task-1199191" [ 1100.111239] env[62368]: _type = "Task" [ 1100.111239] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.119635] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.135564] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1100.135748] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.203s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.181303] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199190, 'name': ReconfigVM_Task, 'duration_secs': 0.158191} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.181639] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1100.186950] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0eb2c7de-b3b2-41a7-b92c-5fcc6c56dad3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.202483] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1100.202483] env[62368]: value = "task-1199192" [ 1100.202483] env[62368]: _type = "Task" [ 1100.202483] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.214135] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199192, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.310715] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.310985] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.399289] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebde1266-29d7-4fde-9670-fe564ae8fabc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.407330] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b470bd-a57b-450c-b5b2-e353455f1e6f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.437052] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2bd879-5097-4705-a2ea-c4ec0c51906d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.443977] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9932ee-c6e4-4966-8b55-2ba2c1215fa1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.458486] env[62368]: DEBUG nova.compute.provider_tree [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.527849] env[62368]: DEBUG nova.objects.instance [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'flavor' on Instance uuid b4070763-a2f2-480f-909c-e323f15a9ac5 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.621544] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199191, 'name': ReconfigVM_Task, 'duration_secs': 0.299654} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.621875] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1/27ff9bed-7c5f-4553-9d8e-8ec78708caf1.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.622528] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6dccebc9-3623-4e88-b58e-e9f9b948422f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.629218] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1100.629218] env[62368]: value = "task-1199193" [ 1100.629218] env[62368]: _type = "Task" [ 1100.629218] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.636634] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199193, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.711483] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199192, 'name': ReconfigVM_Task, 'duration_secs': 0.117004} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.711816] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259906', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'name': 'volume-a5a81572-db8f-4766-b46a-5e14fa56e64c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753bbfe2-405e-4e81-8756-9f52cb637969', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5a81572-db8f-4766-b46a-5e14fa56e64c', 'serial': 'a5a81572-db8f-4766-b46a-5e14fa56e64c'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1100.712151] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1100.712940] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04a34dc-7ab3-44d1-8799-6f1f8d3cc03b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.715684] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.715935] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.716147] env[62368]: INFO nova.compute.manager [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Shelving [ 1100.721148] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1100.722326] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43b420b9-dcf5-49e8-b70b-12802c33f47f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.793184] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1100.793426] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1100.793616] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleting the datastore file [datastore2] 753bbfe2-405e-4e81-8756-9f52cb637969 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1100.793892] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efb8d5e2-f2e0-4266-905d-e064e61fcec3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.799942] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1100.799942] env[62368]: value = "task-1199195" [ 1100.799942] env[62368]: _type = "Task" [ 1100.799942] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.808036] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199195, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.885339] env[62368]: DEBUG oslo_concurrency.lockutils [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "753bbfe2-405e-4e81-8756-9f52cb637969" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.961749] env[62368]: DEBUG nova.scheduler.client.report [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.033695] env[62368]: DEBUG oslo_concurrency.lockutils [None req-6b5fa2ed-cbad-4c06-9b1e-c263bc21d198 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.266s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.138897] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199193, 'name': Rename_Task, 'duration_secs': 0.141462} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.139209] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1101.139454] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8dbdea18-f99e-454d-a3fa-3837847d14fe {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.146363] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1101.146363] env[62368]: value = "task-1199196" [ 1101.146363] env[62368]: _type = "Task" [ 1101.146363] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.158289] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199196, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.223425] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1101.223729] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c43b7ce-5020-436c-a29b-a3f654fd0ebf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.230213] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1101.230213] env[62368]: value = "task-1199197" [ 1101.230213] env[62368]: _type = "Task" [ 1101.230213] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.240040] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199197, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.310140] env[62368]: DEBUG oslo_vmware.api [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199195, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.073296} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.310497] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1101.310702] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1101.310895] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1101.311101] env[62368]: INFO nova.compute.manager [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Took 1.74 seconds to destroy the instance on the hypervisor. [ 1101.311380] env[62368]: DEBUG oslo.service.loopingcall [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1101.311591] env[62368]: DEBUG nova.compute.manager [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1101.311687] env[62368]: DEBUG nova.network.neutron [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1101.466723] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.155s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.467084] env[62368]: INFO nova.compute.manager [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Successfully reverted task state from resize_migrating on failure for instance. [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server [None req-fb99f60b-7212-46f0-b82f-7b5850f7262d tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server raise self.value [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server raise self.value [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server raise self.value [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server raise self.value [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1101.479873] env[62368]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1101.481463] env[62368]: ERROR oslo_messaging.rpc.server [ 1101.656372] env[62368]: DEBUG oslo_vmware.api [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199196, 'name': PowerOnVM_Task, 'duration_secs': 0.441498} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.656633] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1101.656851] env[62368]: INFO nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Took 6.66 seconds to spawn the instance on the hypervisor. [ 1101.657081] env[62368]: DEBUG nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.657854] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f3991b-a25c-440a-83ee-66bf48c80c56 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.740927] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199197, 'name': PowerOffVM_Task, 'duration_secs': 0.245331} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.741216] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1101.742019] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69c5d88-c436-446f-b994-dc78ad9fe30b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.761183] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c214d837-d51d-4947-b158-1f6d0ff0a8a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.930483] env[62368]: DEBUG nova.compute.manager [req-6eea6688-009a-4d4c-b709-5378c9aaf27a req-45de7f30-3db2-4d47-b589-aa6689e65d2d service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Received event network-vif-deleted-11de7a9b-926a-47f0-a28d-a30419725a7f {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.930729] env[62368]: INFO nova.compute.manager [req-6eea6688-009a-4d4c-b709-5378c9aaf27a req-45de7f30-3db2-4d47-b589-aa6689e65d2d service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Neutron deleted interface 11de7a9b-926a-47f0-a28d-a30419725a7f; detaching it from the instance and deleting it from the info cache [ 1101.930923] env[62368]: DEBUG nova.network.neutron [req-6eea6688-009a-4d4c-b709-5378c9aaf27a req-45de7f30-3db2-4d47-b589-aa6689e65d2d service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.174548] env[62368]: INFO nova.compute.manager [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Took 11.60 seconds to build instance. [ 1102.176052] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "f5a33613-790b-4324-9b88-1c39f7c4269c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.176370] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "f5a33613-790b-4324-9b88-1c39f7c4269c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.176681] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "f5a33613-790b-4324-9b88-1c39f7c4269c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.176960] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "f5a33613-790b-4324-9b88-1c39f7c4269c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.177195] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "f5a33613-790b-4324-9b88-1c39f7c4269c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.178822] env[62368]: INFO nova.compute.manager [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Terminating instance [ 1102.180474] env[62368]: DEBUG nova.compute.manager [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1102.180766] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1102.181610] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f5750e-d69b-466a-84a7-c23f9a88ed43 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.190971] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1102.191482] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c66560de-ac71-4bb4-a191-e49748a803c1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.198125] env[62368]: DEBUG oslo_vmware.api [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1102.198125] env[62368]: value = "task-1199198" [ 1102.198125] env[62368]: _type = "Task" [ 1102.198125] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.206503] env[62368]: DEBUG oslo_vmware.api [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199198, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.272788] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1102.273256] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a5f21b3a-9493-4aaa-aaa0-642a02b552bb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.282717] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1102.282717] env[62368]: value = "task-1199199" [ 1102.282717] env[62368]: _type = "Task" [ 1102.282717] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.294612] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199199, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.337742] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.338172] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.403394] env[62368]: DEBUG nova.network.neutron [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.433912] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-26b154cf-bece-4aa9-bac6-c430e0fc3e3b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.443781] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0eb042b-9835-42a1-9bc2-ede239d4f7b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.471459] env[62368]: DEBUG nova.compute.manager [req-6eea6688-009a-4d4c-b709-5378c9aaf27a req-45de7f30-3db2-4d47-b589-aa6689e65d2d service nova] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Detach interface failed, port_id=11de7a9b-926a-47f0-a28d-a30419725a7f, reason: Instance 753bbfe2-405e-4e81-8756-9f52cb637969 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1102.677023] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2da5f635-f8b5-435e-9ac3-84abfbd5bb2e tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.109s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.712028] env[62368]: DEBUG oslo_vmware.api [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199198, 'name': PowerOffVM_Task, 'duration_secs': 0.234047} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.712028] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1102.712028] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1102.712028] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1db0baed-f9bc-4844-8643-11540f6f5621 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.780487] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1102.780487] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1102.780568] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleting the datastore file [datastore2] f5a33613-790b-4324-9b88-1c39f7c4269c {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.780895] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60190d6e-23e4-46ce-98e6-539df2b8323e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.787960] env[62368]: DEBUG oslo_vmware.api [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1102.787960] env[62368]: value = "task-1199201" [ 1102.787960] env[62368]: _type = "Task" [ 1102.787960] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.799131] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199199, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.809709] env[62368]: DEBUG oslo_vmware.api [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199201, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.840823] env[62368]: DEBUG nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1102.906137] env[62368]: INFO nova.compute.manager [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Took 1.59 seconds to deallocate network for instance. [ 1103.294768] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199199, 'name': CreateSnapshot_Task, 'duration_secs': 0.771597} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.295416] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1103.296147] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0513cd02-42d0-463a-95ea-86753040284f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.301090] env[62368]: DEBUG oslo_vmware.api [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199201, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144278} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.301632] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.301822] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1103.302022] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1103.302205] env[62368]: INFO nova.compute.manager [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1103.302448] env[62368]: DEBUG oslo.service.loopingcall [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.302647] env[62368]: DEBUG nova.compute.manager [-] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1103.302740] env[62368]: DEBUG nova.network.neutron [-] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1103.366406] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.366728] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.368836] env[62368]: INFO nova.compute.claims [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1103.452618] env[62368]: INFO nova.compute.manager [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Took 0.55 seconds to detach 1 volumes for instance. [ 1103.454874] env[62368]: DEBUG nova.compute.manager [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Deleting volume: a5a81572-db8f-4766-b46a-5e14fa56e64c {{(pid=62368) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1103.817936] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1103.817936] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-424acd97-97db-4b35-9c20-4cead89589f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.826273] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1103.826273] env[62368]: value = "task-1199203" [ 1103.826273] env[62368]: _type = "Task" [ 1103.826273] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.834379] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199203, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.959955] env[62368]: DEBUG nova.compute.manager [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received event network-changed-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.960335] env[62368]: DEBUG nova.compute.manager [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Refreshing instance network info cache due to event network-changed-7e8eb12d-c67a-4b59-9e66-b3b530e66e75. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1103.960569] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] Acquiring lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.960720] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] Acquired lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.960940] env[62368]: DEBUG nova.network.neutron [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Refreshing network info cache for port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1103.995998] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.038852] env[62368]: DEBUG nova.network.neutron [-] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.338090] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199203, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.496207] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9614ee2-5753-4802-85fd-9a5025fe58a4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.505826] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c965faf-6ff2-4e61-af22-c20321a04543 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.544880] env[62368]: INFO nova.compute.manager [-] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Took 1.24 seconds to deallocate network for instance. [ 1104.547687] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef8d368-a8a9-4f19-967a-69c9a41f5162 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.562315] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f8233b-bd20-43b6-b826-9b3b73905f05 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.578424] env[62368]: DEBUG nova.compute.provider_tree [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.719667] env[62368]: DEBUG nova.network.neutron [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updated VIF entry in instance network info cache for port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1104.720075] env[62368]: DEBUG nova.network.neutron [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [{"id": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "address": "fa:16:3e:0c:7d:cb", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8eb12d-c6", "ovs_interfaceid": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.838569] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199203, 'name': CloneVM_Task, 'duration_secs': 0.969754} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.838879] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Created linked-clone VM from snapshot [ 1104.839772] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767f92ed-6fe9-4050-8ac7-0630bb2b4f8a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.847458] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Uploading image 8950e926-3a83-4e3d-947d-a2a14acfb6e9 {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1104.872501] env[62368]: DEBUG oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1104.872501] env[62368]: value = "vm-259914" [ 1104.872501] env[62368]: _type = "VirtualMachine" [ 1104.872501] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1104.872766] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-903d4a68-84d2-442e-99d7-0a2bcb41906f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.880449] env[62368]: DEBUG oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease: (returnval){ [ 1104.880449] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fe17a8-d98c-570c-2d51-b72897cba0cb" [ 1104.880449] env[62368]: _type = "HttpNfcLease" [ 1104.880449] env[62368]: } obtained for exporting VM: (result){ [ 1104.880449] env[62368]: value = "vm-259914" [ 1104.880449] env[62368]: _type = "VirtualMachine" [ 1104.880449] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1104.880749] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the lease: (returnval){ [ 1104.880749] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fe17a8-d98c-570c-2d51-b72897cba0cb" [ 1104.880749] env[62368]: _type = "HttpNfcLease" [ 1104.880749] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1104.886836] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1104.886836] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fe17a8-d98c-570c-2d51-b72897cba0cb" [ 1104.886836] env[62368]: _type = "HttpNfcLease" [ 1104.886836] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1105.055421] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.081640] env[62368]: DEBUG nova.scheduler.client.report [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1105.222915] env[62368]: DEBUG oslo_concurrency.lockutils [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] Releasing lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.223215] env[62368]: DEBUG nova.compute.manager [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Received event network-vif-deleted-fcd67d36-e1e8-432b-a562-c666ae0c48bf {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.223433] env[62368]: INFO nova.compute.manager [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Neutron deleted interface fcd67d36-e1e8-432b-a562-c666ae0c48bf; detaching it from the instance and deleting it from the info cache [ 1105.223617] env[62368]: DEBUG nova.network.neutron [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.388933] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1105.388933] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fe17a8-d98c-570c-2d51-b72897cba0cb" [ 1105.388933] env[62368]: _type = "HttpNfcLease" [ 1105.388933] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1105.389366] env[62368]: DEBUG oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1105.389366] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fe17a8-d98c-570c-2d51-b72897cba0cb" [ 1105.389366] env[62368]: _type = "HttpNfcLease" [ 1105.389366] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1105.389989] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da510f6-a539-4018-be1a-a0b0b7c51235 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.397168] env[62368]: DEBUG oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b5ca2d-b4db-c21d-a522-9c9422f90e8c/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1105.397357] env[62368]: DEBUG oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b5ca2d-b4db-c21d-a522-9c9422f90e8c/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1105.486434] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4abed376-ad70-4789-9aff-d2af8d07ded0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.586948] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.220s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.587602] env[62368]: DEBUG nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1105.590319] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.594s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.591114] env[62368]: DEBUG nova.objects.instance [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lazy-loading 'resources' on Instance uuid 753bbfe2-405e-4e81-8756-9f52cb637969 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.726966] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b59d0d2b-5437-4305-84df-f1271e754b6b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.736800] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6aeda6-fdb3-483d-9a8c-cce701fc7c19 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.767159] env[62368]: DEBUG nova.compute.manager [req-fe3add18-3411-4520-aa58-3e62850f1046 req-6dd85ba9-c133-407b-a2f8-cf3765df4a13 service nova] [instance: f5a33613-790b-4324-9b88-1c39f7c4269c] Detach interface failed, port_id=fcd67d36-e1e8-432b-a562-c666ae0c48bf, reason: Instance f5a33613-790b-4324-9b88-1c39f7c4269c could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1106.094229] env[62368]: DEBUG nova.compute.utils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1106.098888] env[62368]: DEBUG nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1106.098997] env[62368]: DEBUG nova.network.neutron [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1106.141259] env[62368]: DEBUG nova.policy [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac2df24b03d147f5a33d67a97bbb5bb2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e41fb7a4acf043df959806368ae94ef9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1106.205290] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99aa9f07-e8f9-4fce-8ec5-85bbb2a11af1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.213506] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f400eff3-d486-4db9-aa0f-b09b25d79189 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.250612] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1035442-ceea-4845-b81f-4eb1fdd30d85 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.259664] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa293623-97c9-449a-ba56-affb14e807c5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.275834] env[62368]: DEBUG nova.compute.provider_tree [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.471327] env[62368]: DEBUG nova.network.neutron [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Successfully created port: cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1106.600090] env[62368]: DEBUG nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1106.779378] env[62368]: DEBUG nova.scheduler.client.report [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1107.285227] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.288517] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.233s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.289106] env[62368]: DEBUG nova.objects.instance [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lazy-loading 'resources' on Instance uuid f5a33613-790b-4324-9b88-1c39f7c4269c {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.311372] env[62368]: INFO nova.scheduler.client.report [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted allocations for instance 753bbfe2-405e-4e81-8756-9f52cb637969 [ 1107.612136] env[62368]: DEBUG nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1107.639268] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.639538] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.639705] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.639895] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1107.640062] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1107.640234] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1107.640537] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1107.640987] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1107.640987] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1107.641143] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1107.641323] env[62368]: DEBUG nova.virt.hardware [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.642220] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3898892b-3ce3-43c2-a2ab-dd2ad6ad79ba {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.651299] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb4be40-6b6c-432f-acdf-40cfe5574d5a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.821481] env[62368]: DEBUG oslo_concurrency.lockutils [None req-2d0d93a5-4135-4524-9ebf-1b512939168f tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.258s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.822289] env[62368]: DEBUG oslo_concurrency.lockutils [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.937s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.822496] env[62368]: DEBUG oslo_concurrency.lockutils [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.822712] env[62368]: DEBUG oslo_concurrency.lockutils [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.822885] env[62368]: DEBUG oslo_concurrency.lockutils [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.827259] env[62368]: INFO nova.compute.manager [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Terminating instance [ 1107.831255] env[62368]: DEBUG nova.compute.manager [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1107.831255] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-442a634e-03f9-45c4-ba0b-456868c2dcad {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.842946] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25941fda-1585-4253-acba-9e370fc212f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.877017] env[62368]: WARNING nova.virt.vmwareapi.driver [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 753bbfe2-405e-4e81-8756-9f52cb637969 could not be found. [ 1107.877418] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1107.878606] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10f8a514-da19-4cfa-81d9-3d94ffa4cd32 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.881465] env[62368]: DEBUG nova.compute.manager [req-896f040b-cc34-4c04-a3e1-cacdd4a1eb03 req-26ee021a-bb46-4b4c-9959-5529226dc6c7 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Received event network-vif-plugged-cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.881694] env[62368]: DEBUG oslo_concurrency.lockutils [req-896f040b-cc34-4c04-a3e1-cacdd4a1eb03 req-26ee021a-bb46-4b4c-9959-5529226dc6c7 service nova] Acquiring lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.881918] env[62368]: DEBUG oslo_concurrency.lockutils [req-896f040b-cc34-4c04-a3e1-cacdd4a1eb03 req-26ee021a-bb46-4b4c-9959-5529226dc6c7 service nova] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.882108] env[62368]: DEBUG oslo_concurrency.lockutils [req-896f040b-cc34-4c04-a3e1-cacdd4a1eb03 req-26ee021a-bb46-4b4c-9959-5529226dc6c7 service nova] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.882373] env[62368]: DEBUG nova.compute.manager [req-896f040b-cc34-4c04-a3e1-cacdd4a1eb03 req-26ee021a-bb46-4b4c-9959-5529226dc6c7 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] No waiting events found dispatching network-vif-plugged-cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1107.882475] env[62368]: WARNING nova.compute.manager [req-896f040b-cc34-4c04-a3e1-cacdd4a1eb03 req-26ee021a-bb46-4b4c-9959-5529226dc6c7 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Received unexpected event network-vif-plugged-cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 for instance with vm_state building and task_state spawning. [ 1107.893375] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad1503d-db6f-454c-890c-75f0d428385a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.919632] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9216553f-c326-47be-8b13-867e7ebb0920 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.930955] env[62368]: WARNING nova.virt.vmwareapi.vmops [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 753bbfe2-405e-4e81-8756-9f52cb637969 could not be found. [ 1107.931213] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1107.931405] env[62368]: INFO nova.compute.manager [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Took 0.10 seconds to destroy the instance on the hypervisor. [ 1107.931700] env[62368]: DEBUG oslo.service.loopingcall [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.932403] env[62368]: DEBUG nova.compute.manager [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1107.932506] env[62368]: DEBUG nova.network.neutron [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1107.938263] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd3e0ea-cb23-4bdb-9147-6d646b821157 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.978517] env[62368]: DEBUG nova.network.neutron [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Successfully updated port: cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1107.982162] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43efd04b-8dfb-49c1-b27e-c77334a3949f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.992302] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8f82f2-4971-4160-bc6d-b16e4492775a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.009337] env[62368]: DEBUG nova.compute.provider_tree [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.485960] env[62368]: DEBUG nova.network.neutron [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.487576] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.487730] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.487860] env[62368]: DEBUG nova.network.neutron [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1108.512258] env[62368]: DEBUG nova.scheduler.client.report [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1108.992777] env[62368]: INFO nova.compute.manager [-] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Took 1.06 seconds to deallocate network for instance. [ 1109.007204] env[62368]: WARNING nova.volume.cinder [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Attachment 3852ace7-918a-4578-9cd0-b4c4ec4a9801 does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = 3852ace7-918a-4578-9cd0-b4c4ec4a9801. (HTTP 404) (Request-ID: req-a1ffb253-c154-4cd0-95f9-43d012c04fbd) [ 1109.007539] env[62368]: INFO nova.compute.manager [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Took 0.01 seconds to detach 1 volumes for instance. [ 1109.009903] env[62368]: DEBUG nova.compute.manager [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Deleting volume: a5a81572-db8f-4766-b46a-5e14fa56e64c {{(pid=62368) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1109.017782] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.024865] env[62368]: DEBUG nova.network.neutron [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1109.028320] env[62368]: WARNING nova.compute.manager [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Failed to delete volume: a5a81572-db8f-4766-b46a-5e14fa56e64c due to Volume a5a81572-db8f-4766-b46a-5e14fa56e64c could not be found.: nova.exception.VolumeNotFound: Volume a5a81572-db8f-4766-b46a-5e14fa56e64c could not be found. [ 1109.043038] env[62368]: INFO nova.scheduler.client.report [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted allocations for instance f5a33613-790b-4324-9b88-1c39f7c4269c [ 1109.183656] env[62368]: DEBUG nova.network.neutron [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Updating instance_info_cache with network_info: [{"id": "cb5acbed-41ed-4e72-b4f5-28aea7aa0d69", "address": "fa:16:3e:85:3d:81", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb5acbed-41", "ovs_interfaceid": "cb5acbed-41ed-4e72-b4f5-28aea7aa0d69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.533621] env[62368]: INFO nova.compute.manager [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 753bbfe2-405e-4e81-8756-9f52cb637969] Instance disappeared during terminate [ 1109.533621] env[62368]: DEBUG oslo_concurrency.lockutils [None req-14bb3725-e395-4eb8-abc2-adcef1290807 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "753bbfe2-405e-4e81-8756-9f52cb637969" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.710s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.553177] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ca48c8e3-2ef8-4c43-bfbc-d31b69674311 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "f5a33613-790b-4324-9b88-1c39f7c4269c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.377s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.686435] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.686845] env[62368]: DEBUG nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Instance network_info: |[{"id": "cb5acbed-41ed-4e72-b4f5-28aea7aa0d69", "address": "fa:16:3e:85:3d:81", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb5acbed-41", "ovs_interfaceid": "cb5acbed-41ed-4e72-b4f5-28aea7aa0d69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1109.687414] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:3d:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb5acbed-41ed-4e72-b4f5-28aea7aa0d69', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1109.695649] env[62368]: DEBUG oslo.service.loopingcall [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.695929] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1109.696583] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17327624-17a0-4252-9ff1-35cfe0ae09f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.718232] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1109.718232] env[62368]: value = "task-1199206" [ 1109.718232] env[62368]: _type = "Task" [ 1109.718232] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.726370] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199206, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.900588] env[62368]: DEBUG nova.compute.manager [req-bb9ef260-2ccf-4da4-a43f-08515e1797df req-91c3967d-aac9-469b-acd7-d65387dc3ed0 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Received event network-changed-cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.900840] env[62368]: DEBUG nova.compute.manager [req-bb9ef260-2ccf-4da4-a43f-08515e1797df req-91c3967d-aac9-469b-acd7-d65387dc3ed0 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Refreshing instance network info cache due to event network-changed-cb5acbed-41ed-4e72-b4f5-28aea7aa0d69. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1109.901194] env[62368]: DEBUG oslo_concurrency.lockutils [req-bb9ef260-2ccf-4da4-a43f-08515e1797df req-91c3967d-aac9-469b-acd7-d65387dc3ed0 service nova] Acquiring lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.901394] env[62368]: DEBUG oslo_concurrency.lockutils [req-bb9ef260-2ccf-4da4-a43f-08515e1797df req-91c3967d-aac9-469b-acd7-d65387dc3ed0 service nova] Acquired lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.901619] env[62368]: DEBUG nova.network.neutron [req-bb9ef260-2ccf-4da4-a43f-08515e1797df req-91c3967d-aac9-469b-acd7-d65387dc3ed0 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Refreshing network info cache for port cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.228764] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199206, 'name': CreateVM_Task, 'duration_secs': 0.341381} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.229198] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1110.229639] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.229847] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.230175] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1110.230453] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1eac6fa6-e492-452c-ac77-32087c5400d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.235989] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1110.235989] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]526af712-966c-9ca3-80bb-e7a36994c66c" [ 1110.235989] env[62368]: _type = "Task" [ 1110.235989] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.243818] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526af712-966c-9ca3-80bb-e7a36994c66c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.658772] env[62368]: DEBUG nova.network.neutron [req-bb9ef260-2ccf-4da4-a43f-08515e1797df req-91c3967d-aac9-469b-acd7-d65387dc3ed0 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Updated VIF entry in instance network info cache for port cb5acbed-41ed-4e72-b4f5-28aea7aa0d69. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.659155] env[62368]: DEBUG nova.network.neutron [req-bb9ef260-2ccf-4da4-a43f-08515e1797df req-91c3967d-aac9-469b-acd7-d65387dc3ed0 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Updating instance_info_cache with network_info: [{"id": "cb5acbed-41ed-4e72-b4f5-28aea7aa0d69", "address": "fa:16:3e:85:3d:81", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb5acbed-41", "ovs_interfaceid": "cb5acbed-41ed-4e72-b4f5-28aea7aa0d69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.746332] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]526af712-966c-9ca3-80bb-e7a36994c66c, 'name': SearchDatastore_Task, 'duration_secs': 0.009746} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.746622] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.746867] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1110.747123] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.747309] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.747500] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.747774] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0b7744f-a6c8-4790-973b-f80ec0618eb1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.756011] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.756210] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1110.757031] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-510060c9-99ab-4790-a949-5919bcb8d018 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.762332] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1110.762332] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fa9507-dc75-b2bd-79a5-54d6820dc19d" [ 1110.762332] env[62368]: _type = "Task" [ 1110.762332] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.769901] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fa9507-dc75-b2bd-79a5-54d6820dc19d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.162380] env[62368]: DEBUG oslo_concurrency.lockutils [req-bb9ef260-2ccf-4da4-a43f-08515e1797df req-91c3967d-aac9-469b-acd7-d65387dc3ed0 service nova] Releasing lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.273560] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fa9507-dc75-b2bd-79a5-54d6820dc19d, 'name': SearchDatastore_Task, 'duration_secs': 0.008138} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.274366] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6868aed0-c858-47ae-90e1-a1778c849654 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.279454] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1111.279454] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5230a9e1-caf9-24ee-c72a-13b413be3e56" [ 1111.279454] env[62368]: _type = "Task" [ 1111.279454] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.286580] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5230a9e1-caf9-24ee-c72a-13b413be3e56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.792417] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5230a9e1-caf9-24ee-c72a-13b413be3e56, 'name': SearchDatastore_Task, 'duration_secs': 0.009459} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.792470] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.792801] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 65e4f112-7cf3-4e21-b232-9c1c87ef1da3/65e4f112-7cf3-4e21-b232-9c1c87ef1da3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1111.793266] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f25ba6b-fd1c-4da6-9fe4-8a4544ceb880 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.801329] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1111.801329] env[62368]: value = "task-1199207" [ 1111.801329] env[62368]: _type = "Task" [ 1111.801329] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.812509] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199207, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.312109] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199207, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.417945} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.312468] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 65e4f112-7cf3-4e21-b232-9c1c87ef1da3/65e4f112-7cf3-4e21-b232-9c1c87ef1da3.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1112.312737] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1112.313054] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4df55ec6-64ec-4f54-a915-0ff748dcba1b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.320620] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1112.320620] env[62368]: value = "task-1199208" [ 1112.320620] env[62368]: _type = "Task" [ 1112.320620] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.328291] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199208, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.730765] env[62368]: DEBUG oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b5ca2d-b4db-c21d-a522-9c9422f90e8c/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1112.731690] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4021e3-dcfb-4219-9363-f58e781981fc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.738271] env[62368]: DEBUG oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b5ca2d-b4db-c21d-a522-9c9422f90e8c/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1112.738459] env[62368]: ERROR oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b5ca2d-b4db-c21d-a522-9c9422f90e8c/disk-0.vmdk due to incomplete transfer. [ 1112.738677] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-84e81eba-d7bf-499f-9871-436e5ca1bc83 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.745632] env[62368]: DEBUG oslo_vmware.rw_handles [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b5ca2d-b4db-c21d-a522-9c9422f90e8c/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1112.745830] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Uploaded image 8950e926-3a83-4e3d-947d-a2a14acfb6e9 to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1112.748181] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1112.748408] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8ea3888d-3f0c-4131-8053-ac3e80aaea8b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.755084] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1112.755084] env[62368]: value = "task-1199209" [ 1112.755084] env[62368]: _type = "Task" [ 1112.755084] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.762790] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199209, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.830908] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199208, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098751} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.831169] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1112.831939] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1e42e9-3b74-4061-a5a3-c133bdaba15a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.854045] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 65e4f112-7cf3-4e21-b232-9c1c87ef1da3/65e4f112-7cf3-4e21-b232-9c1c87ef1da3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.854246] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7884f894-cc06-4315-aaba-9745a095d1b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.874080] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1112.874080] env[62368]: value = "task-1199210" [ 1112.874080] env[62368]: _type = "Task" [ 1112.874080] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.881727] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199210, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.983531] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "2a46c954-449e-4d62-be80-add1040ed4c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.983820] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "2a46c954-449e-4d62-be80-add1040ed4c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.984050] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "2a46c954-449e-4d62-be80-add1040ed4c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.984251] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "2a46c954-449e-4d62-be80-add1040ed4c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.984425] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "2a46c954-449e-4d62-be80-add1040ed4c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.986712] env[62368]: INFO nova.compute.manager [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Terminating instance [ 1112.988558] env[62368]: DEBUG nova.compute.manager [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1112.988756] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1112.989588] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b9ff46-ad92-40f6-be0c-688b21588ba1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.997519] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1112.997743] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3fd4b95-cbf1-47f6-a1bc-2fa1daf2b425 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.004309] env[62368]: DEBUG oslo_vmware.api [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1113.004309] env[62368]: value = "task-1199211" [ 1113.004309] env[62368]: _type = "Task" [ 1113.004309] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.012120] env[62368]: DEBUG oslo_vmware.api [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199211, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.266385] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199209, 'name': Destroy_Task} progress is 33%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.383833] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199210, 'name': ReconfigVM_Task, 'duration_secs': 0.264794} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.384273] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 65e4f112-7cf3-4e21-b232-9c1c87ef1da3/65e4f112-7cf3-4e21-b232-9c1c87ef1da3.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.384791] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fd27c0d-db15-4742-a191-3b7e23bb0f97 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.390957] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1113.390957] env[62368]: value = "task-1199212" [ 1113.390957] env[62368]: _type = "Task" [ 1113.390957] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.397952] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199212, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.514769] env[62368]: DEBUG oslo_vmware.api [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199211, 'name': PowerOffVM_Task, 'duration_secs': 0.28746} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.515107] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1113.515289] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1113.515542] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d62b35f-e85d-4042-86a1-ebc9564df62d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.577526] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1113.577787] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1113.577977] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleting the datastore file [datastore1] 2a46c954-449e-4d62-be80-add1040ed4c0 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.578284] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b340bd83-fad4-4fba-a1f3-3688d0c26259 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.586273] env[62368]: DEBUG oslo_vmware.api [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for the task: (returnval){ [ 1113.586273] env[62368]: value = "task-1199214" [ 1113.586273] env[62368]: _type = "Task" [ 1113.586273] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.593399] env[62368]: DEBUG oslo_vmware.api [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199214, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.765540] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199209, 'name': Destroy_Task, 'duration_secs': 0.585276} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.765809] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Destroyed the VM [ 1113.766143] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1113.766423] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0a25337e-1fca-469b-80b5-2c04f8322cc6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.773616] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1113.773616] env[62368]: value = "task-1199215" [ 1113.773616] env[62368]: _type = "Task" [ 1113.773616] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.780992] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199215, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.900060] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199212, 'name': Rename_Task, 'duration_secs': 0.17932} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.900353] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1113.900612] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e950e575-a7f2-4905-b5a0-b4dae83e30d8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.906480] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1113.906480] env[62368]: value = "task-1199216" [ 1113.906480] env[62368]: _type = "Task" [ 1113.906480] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.913685] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199216, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.096494] env[62368]: DEBUG oslo_vmware.api [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Task: {'id': task-1199214, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126322} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.096777] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1114.096971] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1114.097179] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1114.097390] env[62368]: INFO nova.compute.manager [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1114.097651] env[62368]: DEBUG oslo.service.loopingcall [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.097861] env[62368]: DEBUG nova.compute.manager [-] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1114.097957] env[62368]: DEBUG nova.network.neutron [-] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1114.283582] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199215, 'name': RemoveSnapshot_Task, 'duration_secs': 0.49931} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.283848] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1114.284147] env[62368]: DEBUG nova.compute.manager [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1114.284914] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b95261-3ac5-473e-9e6a-7238bca2b8d4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.376159] env[62368]: DEBUG nova.compute.manager [req-3750c3f7-c77d-4dc1-9225-99bac454db1b req-201a2f11-aacd-4f71-bc57-ab36d2e42bd4 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Received event network-vif-deleted-b4fab9b2-299a-4e59-92bf-efdc3d583416 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1114.376159] env[62368]: INFO nova.compute.manager [req-3750c3f7-c77d-4dc1-9225-99bac454db1b req-201a2f11-aacd-4f71-bc57-ab36d2e42bd4 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Neutron deleted interface b4fab9b2-299a-4e59-92bf-efdc3d583416; detaching it from the instance and deleting it from the info cache [ 1114.376159] env[62368]: DEBUG nova.network.neutron [req-3750c3f7-c77d-4dc1-9225-99bac454db1b req-201a2f11-aacd-4f71-bc57-ab36d2e42bd4 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.417671] env[62368]: DEBUG oslo_vmware.api [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199216, 'name': PowerOnVM_Task, 'duration_secs': 0.454935} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.418177] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1114.418378] env[62368]: INFO nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Took 6.81 seconds to spawn the instance on the hypervisor. [ 1114.418566] env[62368]: DEBUG nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1114.419338] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01eb21fa-f554-47f1-90c9-9b6911d7cf63 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.798038] env[62368]: INFO nova.compute.manager [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Shelve offloading [ 1114.799839] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1114.800103] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ff1c74d-4bb5-4689-b0db-8d376a287485 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.808218] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1114.808218] env[62368]: value = "task-1199217" [ 1114.808218] env[62368]: _type = "Task" [ 1114.808218] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.816555] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199217, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.852992] env[62368]: DEBUG nova.network.neutron [-] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.878100] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4bef560e-7957-4045-8ef8-e4758cb24188 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.887333] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f40135-ffd7-4753-bca3-680720789dec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.915583] env[62368]: DEBUG nova.compute.manager [req-3750c3f7-c77d-4dc1-9225-99bac454db1b req-201a2f11-aacd-4f71-bc57-ab36d2e42bd4 service nova] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Detach interface failed, port_id=b4fab9b2-299a-4e59-92bf-efdc3d583416, reason: Instance 2a46c954-449e-4d62-be80-add1040ed4c0 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1114.936100] env[62368]: INFO nova.compute.manager [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Took 11.59 seconds to build instance. [ 1115.291403] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1115.320013] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1115.320233] env[62368]: DEBUG nova.compute.manager [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.320982] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b251102-b174-41b0-91a6-a6b1c21b11c7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.326867] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.327092] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.327310] env[62368]: DEBUG nova.network.neutron [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1115.355042] env[62368]: INFO nova.compute.manager [-] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] Took 1.26 seconds to deallocate network for instance. [ 1115.438276] env[62368]: DEBUG oslo_concurrency.lockutils [None req-48aba15b-e57d-418a-9004-58e4886b2710 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.100s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.702970] env[62368]: DEBUG nova.compute.manager [req-bbfbc142-adba-4200-90ae-724754eab022 req-aaa546d5-d731-447b-9884-49e1dcc49cc8 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Received event network-changed-cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.703681] env[62368]: DEBUG nova.compute.manager [req-bbfbc142-adba-4200-90ae-724754eab022 req-aaa546d5-d731-447b-9884-49e1dcc49cc8 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Refreshing instance network info cache due to event network-changed-cb5acbed-41ed-4e72-b4f5-28aea7aa0d69. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1115.703981] env[62368]: DEBUG oslo_concurrency.lockutils [req-bbfbc142-adba-4200-90ae-724754eab022 req-aaa546d5-d731-447b-9884-49e1dcc49cc8 service nova] Acquiring lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.704186] env[62368]: DEBUG oslo_concurrency.lockutils [req-bbfbc142-adba-4200-90ae-724754eab022 req-aaa546d5-d731-447b-9884-49e1dcc49cc8 service nova] Acquired lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.704368] env[62368]: DEBUG nova.network.neutron [req-bbfbc142-adba-4200-90ae-724754eab022 req-aaa546d5-d731-447b-9884-49e1dcc49cc8 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Refreshing network info cache for port cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1115.795610] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Getting list of instances from cluster (obj){ [ 1115.795610] env[62368]: value = "domain-c8" [ 1115.795610] env[62368]: _type = "ClusterComputeResource" [ 1115.795610] env[62368]: } {{(pid=62368) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1115.796704] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47416e2e-bb1a-45aa-86bb-94fd1be99524 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.810976] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Got total of 4 instances {{(pid=62368) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1115.811197] env[62368]: WARNING nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] While synchronizing instance power states, found 5 instances in the database and 4 instances on the hypervisor. [ 1115.811297] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Triggering sync for uuid 2a46c954-449e-4d62-be80-add1040ed4c0 {{(pid=62368) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1115.811490] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Triggering sync for uuid b4070763-a2f2-480f-909c-e323f15a9ac5 {{(pid=62368) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1115.811646] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Triggering sync for uuid a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 {{(pid=62368) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1115.811795] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Triggering sync for uuid 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 {{(pid=62368) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1115.811945] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Triggering sync for uuid 65e4f112-7cf3-4e21-b232-9c1c87ef1da3 {{(pid=62368) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1115.812315] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "2a46c954-449e-4d62-be80-add1040ed4c0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.812576] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "b4070763-a2f2-480f-909c-e323f15a9ac5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.812770] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.813043] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.813262] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.813448] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.813680] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.813862] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.814712] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe4ba73-00fe-4aef-82c8-d4ade2b56086 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.818077] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5428a1b6-2f93-447b-af91-534cafea1d72 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.820955] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b22bc8-ac8e-4c74-8eaa-2b94beb40008 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.860925] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.861230] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.861453] env[62368]: DEBUG nova.objects.instance [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lazy-loading 'resources' on Instance uuid 2a46c954-449e-4d62-be80-add1040ed4c0 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.047539] env[62368]: DEBUG nova.network.neutron [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [{"id": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "address": "fa:16:3e:58:06:cd", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f003f-8c", "ovs_interfaceid": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.337175] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.524s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.337622] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.524s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.340075] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.527s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.435297] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1e23df-f2d9-419b-9892-c01805e4f343 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.442906] env[62368]: DEBUG nova.network.neutron [req-bbfbc142-adba-4200-90ae-724754eab022 req-aaa546d5-d731-447b-9884-49e1dcc49cc8 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Updated VIF entry in instance network info cache for port cb5acbed-41ed-4e72-b4f5-28aea7aa0d69. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1116.443282] env[62368]: DEBUG nova.network.neutron [req-bbfbc142-adba-4200-90ae-724754eab022 req-aaa546d5-d731-447b-9884-49e1dcc49cc8 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Updating instance_info_cache with network_info: [{"id": "cb5acbed-41ed-4e72-b4f5-28aea7aa0d69", "address": "fa:16:3e:85:3d:81", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb5acbed-41", "ovs_interfaceid": "cb5acbed-41ed-4e72-b4f5-28aea7aa0d69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.445351] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25bef0e-81ff-4665-98c6-5aafa10e876c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.476036] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5000df7d-159a-4931-a4e5-dc8f33ffbbc5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.483885] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9527ec23-a499-4e68-bc42-8f618ff51666 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.497077] env[62368]: DEBUG nova.compute.provider_tree [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.550588] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.773600] env[62368]: DEBUG nova.compute.manager [req-abd4a9b3-a108-4473-a37b-eacc7521a3c9 req-923e0985-a6dd-4880-a7e4-b7c6c93224d3 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received event network-vif-unplugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1116.773600] env[62368]: DEBUG oslo_concurrency.lockutils [req-abd4a9b3-a108-4473-a37b-eacc7521a3c9 req-923e0985-a6dd-4880-a7e4-b7c6c93224d3 service nova] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.773600] env[62368]: DEBUG oslo_concurrency.lockutils [req-abd4a9b3-a108-4473-a37b-eacc7521a3c9 req-923e0985-a6dd-4880-a7e4-b7c6c93224d3 service nova] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.773600] env[62368]: DEBUG oslo_concurrency.lockutils [req-abd4a9b3-a108-4473-a37b-eacc7521a3c9 req-923e0985-a6dd-4880-a7e4-b7c6c93224d3 service nova] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.773600] env[62368]: DEBUG nova.compute.manager [req-abd4a9b3-a108-4473-a37b-eacc7521a3c9 req-923e0985-a6dd-4880-a7e4-b7c6c93224d3 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] No waiting events found dispatching network-vif-unplugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1116.773936] env[62368]: WARNING nova.compute.manager [req-abd4a9b3-a108-4473-a37b-eacc7521a3c9 req-923e0985-a6dd-4880-a7e4-b7c6c93224d3 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received unexpected event network-vif-unplugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 for instance with vm_state shelved and task_state shelving_offloading. [ 1116.863635] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1116.864606] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60cc977-e8ec-4ca3-a6de-b1269325029b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.878052] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1116.878378] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb30738b-285c-4839-b297-476b2cfdaf21 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.949188] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1116.949419] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1116.950075] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleting the datastore file [datastore1] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.950226] env[62368]: DEBUG oslo_concurrency.lockutils [req-bbfbc142-adba-4200-90ae-724754eab022 req-aaa546d5-d731-447b-9884-49e1dcc49cc8 service nova] Releasing lock "refresh_cache-65e4f112-7cf3-4e21-b232-9c1c87ef1da3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.950566] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4aaf4e7f-b03d-4261-b8b3-e46b5b9dcf0b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.958799] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1116.958799] env[62368]: value = "task-1199219" [ 1116.958799] env[62368]: _type = "Task" [ 1116.958799] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.967365] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199219, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.999808] env[62368]: DEBUG nova.scheduler.client.report [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.469786] env[62368]: DEBUG oslo_vmware.api [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199219, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134152} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.470196] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.470326] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1117.470506] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1117.487537] env[62368]: INFO nova.scheduler.client.report [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted allocations for instance a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 [ 1117.506029] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.644s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.522880] env[62368]: INFO nova.scheduler.client.report [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Deleted allocations for instance 2a46c954-449e-4d62-be80-add1040ed4c0 [ 1117.992916] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.992916] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.993107] env[62368]: DEBUG nova.objects.instance [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'resources' on Instance uuid a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.032509] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4e356b45-5b3a-4156-a9ef-eb4d6040e511 tempest-ServerActionsTestOtherA-1622640309 tempest-ServerActionsTestOtherA-1622640309-project-member] Lock "2a46c954-449e-4d62-be80-add1040ed4c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.048s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.033773] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "2a46c954-449e-4d62-be80-add1040ed4c0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.221s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.034102] env[62368]: INFO nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 2a46c954-449e-4d62-be80-add1040ed4c0] During sync_power_state the instance has a pending task (deleting). Skip. [ 1118.034407] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "2a46c954-449e-4d62-be80-add1040ed4c0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.495952] env[62368]: DEBUG nova.objects.instance [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'numa_topology' on Instance uuid a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.798617] env[62368]: DEBUG nova.compute.manager [req-d1af6cca-2930-4ed9-87ed-abdbc3acc283 req-3a35df81-f4f8-4c09-88af-e334162118bc service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received event network-changed-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.798783] env[62368]: DEBUG nova.compute.manager [req-d1af6cca-2930-4ed9-87ed-abdbc3acc283 req-3a35df81-f4f8-4c09-88af-e334162118bc service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Refreshing instance network info cache due to event network-changed-559f003f-8c7a-488b-ae15-6cfa36d5ead5. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1118.799028] env[62368]: DEBUG oslo_concurrency.lockutils [req-d1af6cca-2930-4ed9-87ed-abdbc3acc283 req-3a35df81-f4f8-4c09-88af-e334162118bc service nova] Acquiring lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.799185] env[62368]: DEBUG oslo_concurrency.lockutils [req-d1af6cca-2930-4ed9-87ed-abdbc3acc283 req-3a35df81-f4f8-4c09-88af-e334162118bc service nova] Acquired lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.799428] env[62368]: DEBUG nova.network.neutron [req-d1af6cca-2930-4ed9-87ed-abdbc3acc283 req-3a35df81-f4f8-4c09-88af-e334162118bc service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Refreshing network info cache for port 559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1119.001021] env[62368]: DEBUG nova.objects.base [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1119.059939] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7eafca-ffe9-48c1-877c-9f91b52bfcd2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.069576] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739f03db-ff4c-4472-b1f4-5025a8eb13f7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.108710] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74412760-6c7e-412d-80fe-c18ab5dfcd68 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.116667] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1640e1-254d-45ef-add9-d07ca7c4ac3e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.132727] env[62368]: DEBUG nova.compute.provider_tree [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.545010] env[62368]: DEBUG nova.network.neutron [req-d1af6cca-2930-4ed9-87ed-abdbc3acc283 req-3a35df81-f4f8-4c09-88af-e334162118bc service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updated VIF entry in instance network info cache for port 559f003f-8c7a-488b-ae15-6cfa36d5ead5. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1119.545391] env[62368]: DEBUG nova.network.neutron [req-d1af6cca-2930-4ed9-87ed-abdbc3acc283 req-3a35df81-f4f8-4c09-88af-e334162118bc service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [{"id": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "address": "fa:16:3e:58:06:cd", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": null, "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap559f003f-8c", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.636333] env[62368]: DEBUG nova.scheduler.client.report [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1119.962998] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.048613] env[62368]: DEBUG oslo_concurrency.lockutils [req-d1af6cca-2930-4ed9-87ed-abdbc3acc283 req-3a35df81-f4f8-4c09-88af-e334162118bc service nova] Releasing lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.141220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.148s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.650220] env[62368]: DEBUG oslo_concurrency.lockutils [None req-9bb0a316-8d13-4e2d-80ca-cb5213c2e8aa tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.933s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.650561] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.837s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.650561] env[62368]: INFO nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] During sync_power_state the instance has a pending task (shelving_offloading). Skip. [ 1120.650667] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.650825] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.688s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.650989] env[62368]: INFO nova.compute.manager [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Unshelving [ 1121.686437] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.686714] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.686932] env[62368]: DEBUG nova.objects.instance [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'pci_requests' on Instance uuid a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.193293] env[62368]: DEBUG nova.objects.instance [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'numa_topology' on Instance uuid a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.696218] env[62368]: INFO nova.compute.claims [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1123.159743] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.160020] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.663276] env[62368]: DEBUG nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1123.776710] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26520c7-e5aa-40e4-ad5b-f5200785d43d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.785378] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44208469-e4fb-4c4e-a057-415d14ea5642 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.815415] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3ebe5d-045e-45e3-bc46-d153ab4088f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.823032] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca6b157-b96b-4345-a876-bf8056503afb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.837501] env[62368]: DEBUG nova.compute.provider_tree [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.183874] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.339728] env[62368]: DEBUG nova.scheduler.client.report [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1124.383186] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.846081] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.158s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.848034] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.664s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.849528] env[62368]: INFO nova.compute.claims [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.878588] env[62368]: INFO nova.network.neutron [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating port 559f003f-8c7a-488b-ae15-6cfa36d5ead5 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1125.859705] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.921385] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292be8d8-b93f-4de2-b291-6a4347bdfd46 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.929395] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf52c1bf-7acb-4ae3-b7a1-2ab3a036e0b5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.959215] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2fef7b-20f5-430b-9fc5-895f2cb50393 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.966288] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d9d496-c440-4280-b054-045e63d56592 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.979242] env[62368]: DEBUG nova.compute.provider_tree [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.260437] env[62368]: DEBUG nova.compute.manager [req-faa31c4e-0c38-4a13-8ed9-31da08ba2b7e req-412a0d34-45ee-4e34-a217-3ba5f2d9e7e4 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received event network-vif-plugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1126.260672] env[62368]: DEBUG oslo_concurrency.lockutils [req-faa31c4e-0c38-4a13-8ed9-31da08ba2b7e req-412a0d34-45ee-4e34-a217-3ba5f2d9e7e4 service nova] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.261009] env[62368]: DEBUG oslo_concurrency.lockutils [req-faa31c4e-0c38-4a13-8ed9-31da08ba2b7e req-412a0d34-45ee-4e34-a217-3ba5f2d9e7e4 service nova] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.261214] env[62368]: DEBUG oslo_concurrency.lockutils [req-faa31c4e-0c38-4a13-8ed9-31da08ba2b7e req-412a0d34-45ee-4e34-a217-3ba5f2d9e7e4 service nova] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.261392] env[62368]: DEBUG nova.compute.manager [req-faa31c4e-0c38-4a13-8ed9-31da08ba2b7e req-412a0d34-45ee-4e34-a217-3ba5f2d9e7e4 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] No waiting events found dispatching network-vif-plugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1126.261561] env[62368]: WARNING nova.compute.manager [req-faa31c4e-0c38-4a13-8ed9-31da08ba2b7e req-412a0d34-45ee-4e34-a217-3ba5f2d9e7e4 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received unexpected event network-vif-plugged-559f003f-8c7a-488b-ae15-6cfa36d5ead5 for instance with vm_state shelved_offloaded and task_state spawning. [ 1126.349440] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.349695] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.349896] env[62368]: DEBUG nova.network.neutron [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1126.483118] env[62368]: DEBUG nova.scheduler.client.report [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.860446] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.987200] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.139s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.987773] env[62368]: DEBUG nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1127.051701] env[62368]: DEBUG nova.network.neutron [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [{"id": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "address": "fa:16:3e:58:06:cd", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f003f-8c", "ovs_interfaceid": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.492417] env[62368]: DEBUG nova.compute.utils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.493912] env[62368]: DEBUG nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1127.494133] env[62368]: DEBUG nova.network.neutron [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1127.530784] env[62368]: DEBUG nova.policy [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5e40d7722ab4e9a91e544ed68318d64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d9046c25044b48830356a82644d1b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1127.554359] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.580477] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c01ac7f9c4f2d96882fcdaa94ff15d0c',container_format='bare',created_at=2024-10-08T23:22:41Z,direct_url=,disk_format='vmdk',id=8950e926-3a83-4e3d-947d-a2a14acfb6e9,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-693877824-shelved',owner='73819709966b4e5f873020b399c49871',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-08T23:22:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1127.580750] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1127.580911] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1127.581115] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1127.581275] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1127.581457] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1127.581689] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1127.581854] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1127.582046] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1127.582218] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1127.582415] env[62368]: DEBUG nova.virt.hardware [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1127.583392] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7b6226-d917-40cf-ad18-0c0e0a72da4a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.592718] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8635d12-f04d-4e32-a827-6f3d2db87f1e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.606590] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:06:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0685bd0b-3dbf-4a06-951c-c6a4726dd4b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '559f003f-8c7a-488b-ae15-6cfa36d5ead5', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.614400] env[62368]: DEBUG oslo.service.loopingcall [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.614744] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1127.615150] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31b83287-692e-4364-ab7b-8362274c48eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.634692] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.634692] env[62368]: value = "task-1199220" [ 1127.634692] env[62368]: _type = "Task" [ 1127.634692] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.642901] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199220, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.773014] env[62368]: DEBUG nova.network.neutron [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Successfully created port: 2b52c97e-6de1-4b5a-9dc4-23348784eacb {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1127.997735] env[62368]: DEBUG nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1128.145048] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199220, 'name': CreateVM_Task, 'duration_secs': 0.316192} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.145272] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1128.145982] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.146174] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.146554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1128.146817] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50260b97-262d-406d-ab60-9356adbfc3cf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.152197] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1128.152197] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5295e7b1-5551-d33d-1587-b6d5dc8d428b" [ 1128.152197] env[62368]: _type = "Task" [ 1128.152197] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.166118] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5295e7b1-5551-d33d-1587-b6d5dc8d428b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.310421] env[62368]: DEBUG nova.compute.manager [req-d9410b3a-e878-47f0-91e0-a5313d2003b7 req-69280801-c6c1-4bfa-a5c0-e627f3fb39eb service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received event network-changed-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1128.311131] env[62368]: DEBUG nova.compute.manager [req-d9410b3a-e878-47f0-91e0-a5313d2003b7 req-69280801-c6c1-4bfa-a5c0-e627f3fb39eb service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Refreshing instance network info cache due to event network-changed-559f003f-8c7a-488b-ae15-6cfa36d5ead5. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1128.311131] env[62368]: DEBUG oslo_concurrency.lockutils [req-d9410b3a-e878-47f0-91e0-a5313d2003b7 req-69280801-c6c1-4bfa-a5c0-e627f3fb39eb service nova] Acquiring lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.311131] env[62368]: DEBUG oslo_concurrency.lockutils [req-d9410b3a-e878-47f0-91e0-a5313d2003b7 req-69280801-c6c1-4bfa-a5c0-e627f3fb39eb service nova] Acquired lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.311261] env[62368]: DEBUG nova.network.neutron [req-d9410b3a-e878-47f0-91e0-a5313d2003b7 req-69280801-c6c1-4bfa-a5c0-e627f3fb39eb service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Refreshing network info cache for port 559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1128.662637] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.662894] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Processing image 8950e926-3a83-4e3d-947d-a2a14acfb6e9 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.663146] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9/8950e926-3a83-4e3d-947d-a2a14acfb6e9.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.663299] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9/8950e926-3a83-4e3d-947d-a2a14acfb6e9.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.663480] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.663731] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2cce01dc-c836-4442-9a26-fb02110d8fb4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.681184] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.681362] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1128.682059] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6373701-ddae-4f5f-b0b7-bde77f4ca82f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.687149] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1128.687149] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52cddbaf-12c6-dca8-3810-4deb11c5c9a5" [ 1128.687149] env[62368]: _type = "Task" [ 1128.687149] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.694403] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cddbaf-12c6-dca8-3810-4deb11c5c9a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.855728] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.859232] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.859356] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1129.007698] env[62368]: DEBUG nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1129.019703] env[62368]: DEBUG nova.network.neutron [req-d9410b3a-e878-47f0-91e0-a5313d2003b7 req-69280801-c6c1-4bfa-a5c0-e627f3fb39eb service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updated VIF entry in instance network info cache for port 559f003f-8c7a-488b-ae15-6cfa36d5ead5. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1129.020148] env[62368]: DEBUG nova.network.neutron [req-d9410b3a-e878-47f0-91e0-a5313d2003b7 req-69280801-c6c1-4bfa-a5c0-e627f3fb39eb service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [{"id": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "address": "fa:16:3e:58:06:cd", "network": {"id": "a54835d2-0149-4b42-affe-753c1f5ec7a5", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-237834604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73819709966b4e5f873020b399c49871", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0685bd0b-3dbf-4a06-951c-c6a4726dd4b0", "external-id": "nsx-vlan-transportzone-661", "segmentation_id": 661, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f003f-8c", "ovs_interfaceid": "559f003f-8c7a-488b-ae15-6cfa36d5ead5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.033282] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1129.033532] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1129.033694] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1129.033881] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1129.034047] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1129.034205] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1129.034426] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1129.037025] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1129.037025] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1129.037025] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1129.037025] env[62368]: DEBUG nova.virt.hardware [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.037025] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb75f546-8c39-4e9d-8913-6c340e856e6a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.044573] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e547b435-2f0e-44b5-bb1a-4550eeefcd15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.197351] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Preparing fetch location {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1129.197654] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Fetch image to [datastore2] OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348/OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348.vmdk {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1129.197845] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Downloading stream optimized image 8950e926-3a83-4e3d-947d-a2a14acfb6e9 to [datastore2] OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348/OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348.vmdk on the data store datastore2 as vApp {{(pid=62368) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1129.198030] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Downloading image file data 8950e926-3a83-4e3d-947d-a2a14acfb6e9 to the ESX as VM named 'OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348' {{(pid=62368) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1129.266175] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1129.266175] env[62368]: value = "resgroup-9" [ 1129.266175] env[62368]: _type = "ResourcePool" [ 1129.266175] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1129.266493] env[62368]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9a340623-6ea1-4b17-b91f-003cb48cf5fd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.287928] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease: (returnval){ [ 1129.287928] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52058f2e-c202-6716-0342-3d357df6e0a8" [ 1129.287928] env[62368]: _type = "HttpNfcLease" [ 1129.287928] env[62368]: } obtained for vApp import into resource pool (val){ [ 1129.287928] env[62368]: value = "resgroup-9" [ 1129.287928] env[62368]: _type = "ResourcePool" [ 1129.287928] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1129.288680] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the lease: (returnval){ [ 1129.288680] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52058f2e-c202-6716-0342-3d357df6e0a8" [ 1129.288680] env[62368]: _type = "HttpNfcLease" [ 1129.288680] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1129.295518] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1129.295518] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52058f2e-c202-6716-0342-3d357df6e0a8" [ 1129.295518] env[62368]: _type = "HttpNfcLease" [ 1129.295518] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1129.372782] env[62368]: DEBUG nova.network.neutron [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Successfully updated port: 2b52c97e-6de1-4b5a-9dc4-23348784eacb {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.522859] env[62368]: DEBUG oslo_concurrency.lockutils [req-d9410b3a-e878-47f0-91e0-a5313d2003b7 req-69280801-c6c1-4bfa-a5c0-e627f3fb39eb service nova] Releasing lock "refresh_cache-a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.796777] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1129.796777] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52058f2e-c202-6716-0342-3d357df6e0a8" [ 1129.796777] env[62368]: _type = "HttpNfcLease" [ 1129.796777] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1129.797151] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1129.797151] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52058f2e-c202-6716-0342-3d357df6e0a8" [ 1129.797151] env[62368]: _type = "HttpNfcLease" [ 1129.797151] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1129.797838] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bc97ec-16b1-4c41-bb0b-3faed6993cd2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.806274] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523f0bc8-288e-003b-260c-cd75291b7822/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1129.806456] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523f0bc8-288e-003b-260c-cd75291b7822/disk-0.vmdk. {{(pid=62368) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1129.872638] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-af0ab39e-d8b5-41dd-bc60-2d2e56b0215e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.875781] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.875948] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.876091] env[62368]: DEBUG nova.network.neutron [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1129.896513] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.896678] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquired lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.896822] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Forcefully refreshing network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1130.337087] env[62368]: DEBUG nova.compute.manager [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Received event network-vif-plugged-2b52c97e-6de1-4b5a-9dc4-23348784eacb {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.337465] env[62368]: DEBUG oslo_concurrency.lockutils [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] Acquiring lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.337919] env[62368]: DEBUG oslo_concurrency.lockutils [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] Lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.337919] env[62368]: DEBUG oslo_concurrency.lockutils [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] Lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.338074] env[62368]: DEBUG nova.compute.manager [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] No waiting events found dispatching network-vif-plugged-2b52c97e-6de1-4b5a-9dc4-23348784eacb {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1130.338289] env[62368]: WARNING nova.compute.manager [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Received unexpected event network-vif-plugged-2b52c97e-6de1-4b5a-9dc4-23348784eacb for instance with vm_state building and task_state spawning. [ 1130.338636] env[62368]: DEBUG nova.compute.manager [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Received event network-changed-2b52c97e-6de1-4b5a-9dc4-23348784eacb {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.338877] env[62368]: DEBUG nova.compute.manager [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Refreshing instance network info cache due to event network-changed-2b52c97e-6de1-4b5a-9dc4-23348784eacb. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1130.339097] env[62368]: DEBUG oslo_concurrency.lockutils [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] Acquiring lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.406673] env[62368]: DEBUG nova.network.neutron [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1130.526493] env[62368]: DEBUG nova.network.neutron [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Updating instance_info_cache with network_info: [{"id": "2b52c97e-6de1-4b5a-9dc4-23348784eacb", "address": "fa:16:3e:67:6a:cd", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b52c97e-6d", "ovs_interfaceid": "2b52c97e-6de1-4b5a-9dc4-23348784eacb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.979285] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Completed reading data from the image iterator. {{(pid=62368) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1130.979556] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523f0bc8-288e-003b-260c-cd75291b7822/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1130.980916] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be515060-9953-4102-8815-311396ae7dbd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.988745] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523f0bc8-288e-003b-260c-cd75291b7822/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1130.988979] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523f0bc8-288e-003b-260c-cd75291b7822/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1130.989247] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a86cf573-bc60-4b1e-b8de-7c8df3f8a839 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.029939] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.030289] env[62368]: DEBUG nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Instance network_info: |[{"id": "2b52c97e-6de1-4b5a-9dc4-23348784eacb", "address": "fa:16:3e:67:6a:cd", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b52c97e-6d", "ovs_interfaceid": "2b52c97e-6de1-4b5a-9dc4-23348784eacb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1131.030610] env[62368]: DEBUG oslo_concurrency.lockutils [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] Acquired lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.030854] env[62368]: DEBUG nova.network.neutron [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Refreshing network info cache for port 2b52c97e-6de1-4b5a-9dc4-23348784eacb {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1131.032147] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:6a:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b52c97e-6de1-4b5a-9dc4-23348784eacb', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1131.040558] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Creating folder: Project (09d9046c25044b48830356a82644d1b8). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1131.043837] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-177f1451-9ea0-4d2e-9870-4a1fe2dc8a11 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.057191] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Created folder: Project (09d9046c25044b48830356a82644d1b8) in parent group-v259706. [ 1131.057446] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Creating folder: Instances. Parent ref: group-v259918. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1131.057703] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4e75a33-89db-45ff-9206-789dbdefedf8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.067700] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Created folder: Instances in parent group-v259918. [ 1131.067921] env[62368]: DEBUG oslo.service.loopingcall [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.068133] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1131.068327] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4b7a110-9f14-441e-9a9c-793b30edced1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.089703] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1131.089703] env[62368]: value = "task-1199224" [ 1131.089703] env[62368]: _type = "Task" [ 1131.089703] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.098039] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199224, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.152342] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updating instance_info_cache with network_info: [{"id": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "address": "fa:16:3e:81:12:31", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbd133bd-21", "ovs_interfaceid": "dbd133bd-2126-425e-9772-cbff9f5b8ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.187766] env[62368]: DEBUG oslo_vmware.rw_handles [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523f0bc8-288e-003b-260c-cd75291b7822/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1131.188012] env[62368]: INFO nova.virt.vmwareapi.images [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Downloaded image file data 8950e926-3a83-4e3d-947d-a2a14acfb6e9 [ 1131.188844] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bc0124-0608-4fac-9ecf-7b028c454cf6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.205113] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4249670d-fd2e-4a6d-b7be-a3ca0e070a96 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.231479] env[62368]: INFO nova.virt.vmwareapi.images [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] The imported VM was unregistered [ 1131.234045] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Caching image {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1131.234287] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Creating directory with path [datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.234637] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3ef5ac4-05f5-4f94-902e-a2a2368d745d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.244952] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Created directory with path [datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.245144] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348/OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348.vmdk to [datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9/8950e926-3a83-4e3d-947d-a2a14acfb6e9.vmdk. {{(pid=62368) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1131.245399] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-362c8f8e-5037-4f39-a71d-7b6c0b15c08f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.253162] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1131.253162] env[62368]: value = "task-1199226" [ 1131.253162] env[62368]: _type = "Task" [ 1131.253162] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.261374] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199226, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.601853] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199224, 'name': CreateVM_Task, 'duration_secs': 0.426002} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.602282] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1131.602805] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.602981] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.603338] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1131.603626] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53a1c7a4-6575-4671-b2c7-6ff4cad01281 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.610252] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1131.610252] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524723e5-d18f-ecd7-31d3-a2b424e28287" [ 1131.610252] env[62368]: _type = "Task" [ 1131.610252] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.619381] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524723e5-d18f-ecd7-31d3-a2b424e28287, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.655157] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Releasing lock "refresh_cache-b4070763-a2f2-480f-909c-e323f15a9ac5" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.655876] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updated the network info_cache for instance {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1131.655876] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.655876] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.758570] env[62368]: DEBUG nova.network.neutron [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Updated VIF entry in instance network info cache for port 2b52c97e-6de1-4b5a-9dc4-23348784eacb. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1131.758967] env[62368]: DEBUG nova.network.neutron [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Updating instance_info_cache with network_info: [{"id": "2b52c97e-6de1-4b5a-9dc4-23348784eacb", "address": "fa:16:3e:67:6a:cd", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b52c97e-6d", "ovs_interfaceid": "2b52c97e-6de1-4b5a-9dc4-23348784eacb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.766241] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199226, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.860357] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.860534] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1132.121755] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524723e5-d18f-ecd7-31d3-a2b424e28287, 'name': SearchDatastore_Task, 'duration_secs': 0.069769} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.122041] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.122287] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1132.122536] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.122691] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.122879] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1132.123183] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e67cd3e-fa84-4be6-937a-a8bd8f101063 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.140435] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1132.140649] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1132.141511] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54e1a63a-0ed9-47ec-8d93-fb663c5e625b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.147767] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1132.147767] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527bbcba-031a-ba44-5453-4386994198e8" [ 1132.147767] env[62368]: _type = "Task" [ 1132.147767] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.156650] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527bbcba-031a-ba44-5453-4386994198e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.261572] env[62368]: DEBUG oslo_concurrency.lockutils [req-b79025d8-485f-41e5-9d6c-bbb3ae1f0056 req-7a24f9dc-4b74-4fbc-af4c-5c80cb78ff9c service nova] Releasing lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.265160] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199226, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.659379] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527bbcba-031a-ba44-5453-4386994198e8, 'name': SearchDatastore_Task, 'duration_secs': 0.079905} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.660226] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d1f4ffb-593a-4498-bc66-adbea1b4863c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.666483] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1132.666483] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5256a89c-9e05-33ef-3b7d-ee6bdbc93471" [ 1132.666483] env[62368]: _type = "Task" [ 1132.666483] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.675271] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5256a89c-9e05-33ef-3b7d-ee6bdbc93471, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.765138] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199226, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.860632] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.177580] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5256a89c-9e05-33ef-3b7d-ee6bdbc93471, 'name': SearchDatastore_Task, 'duration_secs': 0.085614} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.177908] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.178355] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70/0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1133.178717] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e887504-c86c-4120-97a8-7d96697e65ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.187366] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1133.187366] env[62368]: value = "task-1199227" [ 1133.187366] env[62368]: _type = "Task" [ 1133.187366] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.196452] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199227, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.265500] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199226, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.363946] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.364250] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.364435] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.364625] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1133.365632] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e29cd4-76c7-41ee-b378-8236da5e37b6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.375650] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d061f3-df66-4e64-87c2-4af3eabf1555 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.393759] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ff6448-1389-4812-8e1e-4f5454e8a5cf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.402051] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba51dc4-c471-4e89-b8b3-d3a8409f2391 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.433848] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180242MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1133.434070] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.434229] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.697819] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199227, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.766045] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199226, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.373816} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.766233] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348/OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348.vmdk to [datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9/8950e926-3a83-4e3d-947d-a2a14acfb6e9.vmdk. [ 1133.766277] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Cleaning up location [datastore2] OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1133.766449] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_a2dbcc8f-e86e-46b8-b4ca-7f46cee38348 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.766703] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-962e293a-132c-4dd7-84df-3f3b495bb1c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.772958] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1133.772958] env[62368]: value = "task-1199228" [ 1133.772958] env[62368]: _type = "Task" [ 1133.772958] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.780972] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199228, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.198839] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199227, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.282747] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199228, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035173} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.283017] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.283191] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9/8950e926-3a83-4e3d-947d-a2a14acfb6e9.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.283444] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9/8950e926-3a83-4e3d-947d-a2a14acfb6e9.vmdk to [datastore2] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2/a1fc23d4-dcbb-4df9-bd5a-74377bee36d2.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1134.283705] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b5209ea-bff9-4e8a-b95c-91f4981a61f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.289937] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1134.289937] env[62368]: value = "task-1199229" [ 1134.289937] env[62368]: _type = "Task" [ 1134.289937] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.297428] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199229, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.460302] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b4070763-a2f2-480f-909c-e323f15a9ac5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.460467] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.460594] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 65e4f112-7cf3-4e21-b232-9c1c87ef1da3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.460711] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.460832] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.461062] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1134.461212] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1134.528834] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91f5030-58c3-4e28-a97e-7d9ef98d52a3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.537627] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dcb79cc-b4f0-49c2-bcbc-68f519054e5e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.567101] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4997166a-70c8-4714-befc-a646b899dcd4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.574708] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f33557-0802-4824-bc6e-e4f5dedd4d49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.588802] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.700417] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199227, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.800146] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199229, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.092265] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1135.201211] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199227, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.799347} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.201546] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70/0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1135.201804] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1135.202129] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dac443b6-c4a0-4a21-a57e-c3e303530a49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.210045] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1135.210045] env[62368]: value = "task-1199230" [ 1135.210045] env[62368]: _type = "Task" [ 1135.210045] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.219585] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199230, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.301903] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199229, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.597057] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1135.597337] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.163s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.720634] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199230, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113622} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.721069] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1135.723143] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5246042c-1efc-441b-a6b7-5a7ec5212e7f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.743768] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70/0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.744123] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1db851b-5fbd-4baa-a00a-b1be946ed6ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.764189] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1135.764189] env[62368]: value = "task-1199231" [ 1135.764189] env[62368]: _type = "Task" [ 1135.764189] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.772598] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199231, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.799419] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199229, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.274048] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199231, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.300162] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199229, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.775878] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199231, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.800814] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199229, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.276622] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199231, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.300714] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199229, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.837465} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.300970] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8950e926-3a83-4e3d-947d-a2a14acfb6e9/8950e926-3a83-4e3d-947d-a2a14acfb6e9.vmdk to [datastore2] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2/a1fc23d4-dcbb-4df9-bd5a-74377bee36d2.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1137.301778] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc65d8f-0389-4ede-aad3-222f03850198 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.323505] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2/a1fc23d4-dcbb-4df9-bd5a-74377bee36d2.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.323757] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4bb345f-7f4d-4530-bb49-b2191a0c5434 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.342446] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1137.342446] env[62368]: value = "task-1199232" [ 1137.342446] env[62368]: _type = "Task" [ 1137.342446] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.349723] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199232, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.776645] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199231, 'name': ReconfigVM_Task, 'duration_secs': 1.864515} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.777032] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70/0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.777590] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e57d8248-f266-427b-b0dc-8f2c844da8a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.783903] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1137.783903] env[62368]: value = "task-1199233" [ 1137.783903] env[62368]: _type = "Task" [ 1137.783903] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.791692] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199233, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.851765] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199232, 'name': ReconfigVM_Task, 'duration_secs': 0.38163} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.852057] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Reconfigured VM instance instance-0000006b to attach disk [datastore2] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2/a1fc23d4-dcbb-4df9-bd5a-74377bee36d2.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.852682] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-946eae06-d739-42b0-9ca9-30b96adaf046 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.858797] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1137.858797] env[62368]: value = "task-1199234" [ 1137.858797] env[62368]: _type = "Task" [ 1137.858797] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.871082] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199234, 'name': Rename_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.293834] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199233, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.368378] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199234, 'name': Rename_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.794059] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199233, 'name': Rename_Task, 'duration_secs': 0.537793} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.794563] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1138.794563] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-145b9f89-3462-44f9-8ed7-1f8c7da5cb76 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.800915] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1138.800915] env[62368]: value = "task-1199235" [ 1138.800915] env[62368]: _type = "Task" [ 1138.800915] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.808683] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199235, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.868823] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199234, 'name': Rename_Task, 'duration_secs': 0.536378} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.869092] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1138.869348] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2da62ce2-917e-4a02-9258-809b2eedc00b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.875130] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1138.875130] env[62368]: value = "task-1199236" [ 1138.875130] env[62368]: _type = "Task" [ 1138.875130] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.882757] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199236, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.311904] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199235, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.385607] env[62368]: DEBUG oslo_vmware.api [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199236, 'name': PowerOnVM_Task, 'duration_secs': 0.474242} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.385916] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1139.486232] env[62368]: DEBUG nova.compute.manager [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1139.487212] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19160f6-9c1b-4f13-96cf-6a035fc9fc80 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.811361] env[62368]: DEBUG oslo_vmware.api [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199235, 'name': PowerOnVM_Task, 'duration_secs': 0.626885} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.811694] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1139.811694] env[62368]: INFO nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Took 10.80 seconds to spawn the instance on the hypervisor. [ 1139.811846] env[62368]: DEBUG nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1139.812781] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96d0bb1-fc51-4bda-8d6e-9be26ac6164f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.002816] env[62368]: DEBUG oslo_concurrency.lockutils [None req-fe061747-acc0-4b35-866c-883836630a82 tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.352s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.331299] env[62368]: INFO nova.compute.manager [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Took 16.16 seconds to build instance. [ 1140.833181] env[62368]: DEBUG oslo_concurrency.lockutils [None req-131ed619-c926-4af2-ad8a-de198909dda7 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.673s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.986443] env[62368]: INFO nova.compute.manager [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Rescuing [ 1140.986727] env[62368]: DEBUG oslo_concurrency.lockutils [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.986897] env[62368]: DEBUG oslo_concurrency.lockutils [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.987121] env[62368]: DEBUG nova.network.neutron [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1141.089294] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.089607] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.089864] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.090108] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.090326] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.092926] env[62368]: INFO nova.compute.manager [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Terminating instance [ 1141.094968] env[62368]: DEBUG nova.compute.manager [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1141.095227] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1141.096173] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33babb58-9d0b-4626-ae23-8f0ac7356f31 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.104510] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1141.104784] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73ead2b6-9c5e-483f-8cdd-7e0fd91894cd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.111387] env[62368]: DEBUG oslo_vmware.api [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1141.111387] env[62368]: value = "task-1199237" [ 1141.111387] env[62368]: _type = "Task" [ 1141.111387] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.119666] env[62368]: DEBUG oslo_vmware.api [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199237, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.620916] env[62368]: DEBUG oslo_vmware.api [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199237, 'name': PowerOffVM_Task, 'duration_secs': 0.232426} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.621209] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1141.621386] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1141.621863] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6f4ff64-56bb-4bca-b308-403ccfbad53e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.685863] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1141.686114] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1141.686290] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleting the datastore file [datastore2] a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1141.686557] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9782b4ca-4dc4-472f-9e48-7e6f89e26453 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.693291] env[62368]: DEBUG oslo_vmware.api [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for the task: (returnval){ [ 1141.693291] env[62368]: value = "task-1199239" [ 1141.693291] env[62368]: _type = "Task" [ 1141.693291] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.700885] env[62368]: DEBUG oslo_vmware.api [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199239, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.721150] env[62368]: DEBUG nova.network.neutron [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Updating instance_info_cache with network_info: [{"id": "2b52c97e-6de1-4b5a-9dc4-23348784eacb", "address": "fa:16:3e:67:6a:cd", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b52c97e-6d", "ovs_interfaceid": "2b52c97e-6de1-4b5a-9dc4-23348784eacb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.953722] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.954132] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.954297] env[62368]: INFO nova.compute.manager [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Shelving [ 1142.204258] env[62368]: DEBUG oslo_vmware.api [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Task: {'id': task-1199239, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149697} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.204473] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1142.204660] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1142.204840] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1142.205025] env[62368]: INFO nova.compute.manager [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1142.205283] env[62368]: DEBUG oslo.service.loopingcall [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1142.205514] env[62368]: DEBUG nova.compute.manager [-] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1142.205617] env[62368]: DEBUG nova.network.neutron [-] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1142.223411] env[62368]: DEBUG oslo_concurrency.lockutils [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "refresh_cache-0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.462084] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1142.462288] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82db6e2c-7c9f-4585-b95a-b036888c4109 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.469775] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1142.469775] env[62368]: value = "task-1199240" [ 1142.469775] env[62368]: _type = "Task" [ 1142.469775] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.478956] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.635624] env[62368]: DEBUG nova.compute.manager [req-faf8da4e-3e9b-435c-8773-5c80ef5e2874 req-084c636a-eb85-4803-aa33-154d414e90f6 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Received event network-vif-deleted-559f003f-8c7a-488b-ae15-6cfa36d5ead5 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1142.635836] env[62368]: INFO nova.compute.manager [req-faf8da4e-3e9b-435c-8773-5c80ef5e2874 req-084c636a-eb85-4803-aa33-154d414e90f6 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Neutron deleted interface 559f003f-8c7a-488b-ae15-6cfa36d5ead5; detaching it from the instance and deleting it from the info cache [ 1142.636099] env[62368]: DEBUG nova.network.neutron [req-faf8da4e-3e9b-435c-8773-5c80ef5e2874 req-084c636a-eb85-4803-aa33-154d414e90f6 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.750022] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1142.750303] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9b72f59-059d-4e4f-bd91-fadc2da23d31 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.758391] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1142.758391] env[62368]: value = "task-1199241" [ 1142.758391] env[62368]: _type = "Task" [ 1142.758391] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.767120] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199241, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.980063] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199240, 'name': PowerOffVM_Task, 'duration_secs': 0.183607} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.980417] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1142.981118] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bfcbc7-1b58-4539-ab26-c1907f7d2d64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.999263] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265508f8-4138-406e-aacd-bc171d2d321e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.102787] env[62368]: DEBUG nova.network.neutron [-] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.140039] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cba535cc-8caf-4b58-8888-3f7a4b6b482a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.152119] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8471af-681d-4f9e-bd30-130d60095124 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.179209] env[62368]: DEBUG nova.compute.manager [req-faf8da4e-3e9b-435c-8773-5c80ef5e2874 req-084c636a-eb85-4803-aa33-154d414e90f6 service nova] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Detach interface failed, port_id=559f003f-8c7a-488b-ae15-6cfa36d5ead5, reason: Instance a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1143.268707] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199241, 'name': PowerOffVM_Task, 'duration_secs': 0.195963} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.268975] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1143.271149] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5319836b-0d1b-47ee-a032-acc608bf2c7c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.289210] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cae510-9399-4e66-aa1f-363e970309ce {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.319295] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1143.319556] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f9e40d9-9e15-42f3-9680-f2697e3df102 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.325505] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1143.325505] env[62368]: value = "task-1199242" [ 1143.325505] env[62368]: _type = "Task" [ 1143.325505] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.332432] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199242, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.509613] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Creating Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1143.509953] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-18eb54ef-3a18-4019-8bf9-14fdf2585a6c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.517769] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1143.517769] env[62368]: value = "task-1199243" [ 1143.517769] env[62368]: _type = "Task" [ 1143.517769] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.527641] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199243, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.605766] env[62368]: INFO nova.compute.manager [-] [instance: a1fc23d4-dcbb-4df9-bd5a-74377bee36d2] Took 1.40 seconds to deallocate network for instance. [ 1143.836186] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1143.836445] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1143.836720] env[62368]: DEBUG oslo_concurrency.lockutils [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.836894] env[62368]: DEBUG oslo_concurrency.lockutils [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.837090] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1143.837357] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85b84a63-90b9-4d79-b9f8-b5ccf969b756 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.846066] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1143.846267] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1143.847014] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6a73034-d4ed-46de-a229-b95ccdc93f6b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.853053] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1143.853053] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]521b8a3e-2049-1212-ea13-a0e715fb8e2b" [ 1143.853053] env[62368]: _type = "Task" [ 1143.853053] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.859982] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521b8a3e-2049-1212-ea13-a0e715fb8e2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.027708] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199243, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.113272] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.113511] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.113791] env[62368]: DEBUG nova.objects.instance [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lazy-loading 'resources' on Instance uuid a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.362743] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]521b8a3e-2049-1212-ea13-a0e715fb8e2b, 'name': SearchDatastore_Task, 'duration_secs': 0.008842} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.363570] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0a96f0b-2464-415e-a8f6-58c57079d725 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.368219] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1144.368219] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52701f6c-6223-d889-85ef-8eb010274d8d" [ 1144.368219] env[62368]: _type = "Task" [ 1144.368219] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.375372] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52701f6c-6223-d889-85ef-8eb010274d8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.527451] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199243, 'name': CreateSnapshot_Task, 'duration_secs': 0.725334} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.528321] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Created Snapshot of the VM instance {{(pid=62368) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1144.529093] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0b95c1-247f-4528-8aad-5733d70f05f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.685744] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e9d670-c2f8-43c6-8581-b779f56957e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.692928] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c34ed0d-8a9f-4979-95fe-508b658aea4f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.723431] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b832a0cc-7e49-4f7b-bc16-f6f191b84634 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.729982] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4ba1f3-dc69-4a5e-891a-8718c7ab1d02 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.742468] env[62368]: DEBUG nova.compute.provider_tree [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1144.877845] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52701f6c-6223-d889-85ef-8eb010274d8d, 'name': SearchDatastore_Task, 'duration_secs': 0.010674} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.878123] env[62368]: DEBUG oslo_concurrency.lockutils [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.878473] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. {{(pid=62368) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1144.879017] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-92dfe373-7ecb-43a9-8266-1e391cc0060d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.885954] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1144.885954] env[62368]: value = "task-1199244" [ 1144.885954] env[62368]: _type = "Task" [ 1144.885954] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.893126] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199244, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.045953] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Creating linked-clone VM from snapshot {{(pid=62368) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1145.046398] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7ca4d25d-3656-4860-96ea-48a84afbb043 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.054725] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1145.054725] env[62368]: value = "task-1199245" [ 1145.054725] env[62368]: _type = "Task" [ 1145.054725] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.062981] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199245, 'name': CloneVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.245732] env[62368]: DEBUG nova.scheduler.client.report [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1145.396755] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199244, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.564687] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199245, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.751123] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.779681] env[62368]: INFO nova.scheduler.client.report [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Deleted allocations for instance a1fc23d4-dcbb-4df9-bd5a-74377bee36d2 [ 1145.896734] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199244, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533616} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.896913] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. [ 1145.897709] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8997410a-81f9-4434-96aa-33b81d79e37d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.922846] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1145.923125] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b1559af-3763-45ae-b611-84242763a4ff {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.941115] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1145.941115] env[62368]: value = "task-1199246" [ 1145.941115] env[62368]: _type = "Task" [ 1145.941115] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.948513] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199246, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.065597] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199245, 'name': CloneVM_Task} progress is 94%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.287784] env[62368]: DEBUG oslo_concurrency.lockutils [None req-79a0386b-b0e9-4d17-9b46-939e5f084abc tempest-ServerActionsTestOtherB-858048377 tempest-ServerActionsTestOtherB-858048377-project-member] Lock "a1fc23d4-dcbb-4df9-bd5a-74377bee36d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.198s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.450540] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199246, 'name': ReconfigVM_Task, 'duration_secs': 0.313561} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.450760] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1146.451582] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d80602-eb1f-4a10-a61a-9b13d4f72e14 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.474992] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19e8b208-0b0f-4aa6-8081-aaa8b100f12b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.489590] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1146.489590] env[62368]: value = "task-1199248" [ 1146.489590] env[62368]: _type = "Task" [ 1146.489590] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.497127] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199248, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.565904] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199245, 'name': CloneVM_Task, 'duration_secs': 1.459641} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.566161] env[62368]: INFO nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Created linked-clone VM from snapshot [ 1146.566895] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3081d88-a6bc-476f-b702-4147b1863c67 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.573748] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Uploading image cf2811ab-2887-48e1-8edf-05289097fce3 {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1146.600239] env[62368]: DEBUG oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1146.600239] env[62368]: value = "vm-259922" [ 1146.600239] env[62368]: _type = "VirtualMachine" [ 1146.600239] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1146.600534] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ae87e2ec-f7f7-4fdf-9464-add8aa7f802d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.606832] env[62368]: DEBUG oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lease: (returnval){ [ 1146.606832] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5210959e-1d83-5ffb-578a-a927f4a6e699" [ 1146.606832] env[62368]: _type = "HttpNfcLease" [ 1146.606832] env[62368]: } obtained for exporting VM: (result){ [ 1146.606832] env[62368]: value = "vm-259922" [ 1146.606832] env[62368]: _type = "VirtualMachine" [ 1146.606832] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1146.607123] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the lease: (returnval){ [ 1146.607123] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5210959e-1d83-5ffb-578a-a927f4a6e699" [ 1146.607123] env[62368]: _type = "HttpNfcLease" [ 1146.607123] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1146.612842] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1146.612842] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5210959e-1d83-5ffb-578a-a927f4a6e699" [ 1146.612842] env[62368]: _type = "HttpNfcLease" [ 1146.612842] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1146.999323] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199248, 'name': ReconfigVM_Task, 'duration_secs': 0.156275} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.999572] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1146.999835] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bad086e-645e-4fc6-bdc6-5c883bd0abef {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.006354] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1147.006354] env[62368]: value = "task-1199250" [ 1147.006354] env[62368]: _type = "Task" [ 1147.006354] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.013609] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199250, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.116048] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1147.116048] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5210959e-1d83-5ffb-578a-a927f4a6e699" [ 1147.116048] env[62368]: _type = "HttpNfcLease" [ 1147.116048] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1147.116048] env[62368]: DEBUG oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1147.116048] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5210959e-1d83-5ffb-578a-a927f4a6e699" [ 1147.116048] env[62368]: _type = "HttpNfcLease" [ 1147.116048] env[62368]: }. {{(pid=62368) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1147.116683] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587459a0-9dc9-4c46-abd9-8ce0e69cf7d3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.124280] env[62368]: DEBUG oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528e7730-444f-e73d-5884-13d45ef403d6/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1147.124511] env[62368]: DEBUG oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528e7730-444f-e73d-5884-13d45ef403d6/disk-0.vmdk for reading. {{(pid=62368) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1147.220566] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8b53143d-8d25-4b64-99b9-810be4055593 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.518171] env[62368]: DEBUG oslo_vmware.api [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199250, 'name': PowerOnVM_Task, 'duration_secs': 0.480729} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.518390] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1147.521321] env[62368]: DEBUG nova.compute.manager [None req-234b2d31-997b-400e-86c0-cfc039a6ad2a tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1147.522124] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2c8ac9-b16c-46a9-917f-449e7b7c0e82 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.406036] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "9b792344-4397-4c8d-906e-64a4c7d91606" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.406558] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "9b792344-4397-4c8d-906e-64a4c7d91606" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.909189] env[62368]: DEBUG nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1150.442439] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.442755] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.445186] env[62368]: INFO nova.compute.claims [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1151.531601] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a71319-41c0-490f-a4f6-b1d311977a36 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.539639] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ae3875-a27a-4f3b-a925-a3f90c420705 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.570736] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9f81c8-f312-4456-9811-41026b487c19 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.579030] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f04a9c-c8c4-4a2b-907d-2d06a89f1723 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.592744] env[62368]: DEBUG nova.compute.provider_tree [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.096112] env[62368]: DEBUG nova.scheduler.client.report [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.601267] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.158s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.601901] env[62368]: DEBUG nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1153.106753] env[62368]: DEBUG nova.compute.utils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1153.108216] env[62368]: DEBUG nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1153.108386] env[62368]: DEBUG nova.network.neutron [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1153.149702] env[62368]: DEBUG nova.policy [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5e40d7722ab4e9a91e544ed68318d64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09d9046c25044b48830356a82644d1b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1153.463487] env[62368]: DEBUG nova.network.neutron [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Successfully created port: 9a09cec0-45b4-46b9-872f-e153d6ea3791 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1153.614595] env[62368]: DEBUG nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1154.105252] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.106133] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.106133] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.106354] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.106474] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.109899] env[62368]: INFO nova.compute.manager [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Terminating instance [ 1154.111256] env[62368]: DEBUG nova.compute.manager [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1154.111256] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1154.112277] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f24f4f-9816-4248-9ef9-e3a4e03b47c0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.124022] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1154.124022] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c30ca646-2558-4abf-8931-89e434eec1b3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.131538] env[62368]: DEBUG oslo_vmware.api [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1154.131538] env[62368]: value = "task-1199251" [ 1154.131538] env[62368]: _type = "Task" [ 1154.131538] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.140448] env[62368]: DEBUG oslo_vmware.api [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.626305] env[62368]: DEBUG nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1154.642319] env[62368]: DEBUG oslo_vmware.api [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199251, 'name': PowerOffVM_Task, 'duration_secs': 0.310937} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.642661] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1154.642876] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1154.643202] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11cd451d-0103-4f2a-9cce-991ebdc0ebd7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.656352] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1154.656850] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1154.657744] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1154.657744] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1154.657910] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1154.657967] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1154.658308] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1154.658521] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1154.658786] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1154.659018] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1154.659259] env[62368]: DEBUG nova.virt.hardware [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1154.660186] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c56508-34c2-4713-bbeb-089c163372f9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.668475] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29005c4-eb22-49bc-8529-e0443d93efb1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.715902] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1154.716219] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1154.716470] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleting the datastore file [datastore2] 65e4f112-7cf3-4e21-b232-9c1c87ef1da3 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.716773] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1d1d5da-09e9-49a8-91b5-b4e257586d84 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.724884] env[62368]: DEBUG oslo_vmware.api [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1154.724884] env[62368]: value = "task-1199253" [ 1154.724884] env[62368]: _type = "Task" [ 1154.724884] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.733866] env[62368]: DEBUG oslo_vmware.api [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199253, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.905850] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquiring lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.906806] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.996105] env[62368]: DEBUG nova.compute.manager [req-f534c100-b540-4c61-85d9-62cef4ec9263 req-1f120547-3f70-4793-8dbf-51fa75f1f120 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Received event network-vif-plugged-9a09cec0-45b4-46b9-872f-e153d6ea3791 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1154.996657] env[62368]: DEBUG oslo_concurrency.lockutils [req-f534c100-b540-4c61-85d9-62cef4ec9263 req-1f120547-3f70-4793-8dbf-51fa75f1f120 service nova] Acquiring lock "9b792344-4397-4c8d-906e-64a4c7d91606-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.996881] env[62368]: DEBUG oslo_concurrency.lockutils [req-f534c100-b540-4c61-85d9-62cef4ec9263 req-1f120547-3f70-4793-8dbf-51fa75f1f120 service nova] Lock "9b792344-4397-4c8d-906e-64a4c7d91606-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.997093] env[62368]: DEBUG oslo_concurrency.lockutils [req-f534c100-b540-4c61-85d9-62cef4ec9263 req-1f120547-3f70-4793-8dbf-51fa75f1f120 service nova] Lock "9b792344-4397-4c8d-906e-64a4c7d91606-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.997269] env[62368]: DEBUG nova.compute.manager [req-f534c100-b540-4c61-85d9-62cef4ec9263 req-1f120547-3f70-4793-8dbf-51fa75f1f120 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] No waiting events found dispatching network-vif-plugged-9a09cec0-45b4-46b9-872f-e153d6ea3791 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1154.997468] env[62368]: WARNING nova.compute.manager [req-f534c100-b540-4c61-85d9-62cef4ec9263 req-1f120547-3f70-4793-8dbf-51fa75f1f120 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Received unexpected event network-vif-plugged-9a09cec0-45b4-46b9-872f-e153d6ea3791 for instance with vm_state building and task_state spawning. [ 1155.023285] env[62368]: DEBUG oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528e7730-444f-e73d-5884-13d45ef403d6/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1155.024444] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40fa563-bcec-4d6c-bf4b-8b7dc4cc1f35 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.031023] env[62368]: DEBUG oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528e7730-444f-e73d-5884-13d45ef403d6/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1155.031222] env[62368]: ERROR oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528e7730-444f-e73d-5884-13d45ef403d6/disk-0.vmdk due to incomplete transfer. [ 1155.031475] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0d98310c-324b-47a5-9241-badd8c3f7651 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.038884] env[62368]: DEBUG oslo_vmware.rw_handles [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528e7730-444f-e73d-5884-13d45ef403d6/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1155.039132] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Uploaded image cf2811ab-2887-48e1-8edf-05289097fce3 to the Glance image server {{(pid=62368) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1155.041550] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Destroying the VM {{(pid=62368) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1155.041797] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-df45e3f1-ee08-452f-8582-d472a7b2793e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.048583] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1155.048583] env[62368]: value = "task-1199254" [ 1155.048583] env[62368]: _type = "Task" [ 1155.048583] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.057593] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199254, 'name': Destroy_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.087321] env[62368]: DEBUG nova.network.neutron [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Successfully updated port: 9a09cec0-45b4-46b9-872f-e153d6ea3791 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1155.236021] env[62368]: DEBUG oslo_vmware.api [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199253, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299353} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.236319] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1155.236528] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1155.236733] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1155.236942] env[62368]: INFO nova.compute.manager [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1155.237244] env[62368]: DEBUG oslo.service.loopingcall [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1155.237477] env[62368]: DEBUG nova.compute.manager [-] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1155.237600] env[62368]: DEBUG nova.network.neutron [-] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1155.409365] env[62368]: DEBUG nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1155.558438] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199254, 'name': Destroy_Task, 'duration_secs': 0.355714} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.558722] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Destroyed the VM [ 1155.558982] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Deleting Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1155.563405] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-510edc7b-f931-40ba-be02-619858bcbe8c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.570347] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1155.570347] env[62368]: value = "task-1199255" [ 1155.570347] env[62368]: _type = "Task" [ 1155.570347] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.580600] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199255, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.590322] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.590462] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.590611] env[62368]: DEBUG nova.network.neutron [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1155.941554] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.941883] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.944354] env[62368]: INFO nova.compute.claims [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1156.080285] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199255, 'name': RemoveSnapshot_Task, 'duration_secs': 0.392975} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.080579] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Deleted Snapshot of the VM instance {{(pid=62368) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1156.080845] env[62368]: DEBUG nova.compute.manager [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1156.081655] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3db3107-296e-4cf1-ac9e-ee4787f2ee8d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.123270] env[62368]: DEBUG nova.network.neutron [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1156.185669] env[62368]: DEBUG nova.network.neutron [-] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.241883] env[62368]: DEBUG nova.network.neutron [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Updating instance_info_cache with network_info: [{"id": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "address": "fa:16:3e:d6:e9:70", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a09cec0-45", "ovs_interfaceid": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.593288] env[62368]: INFO nova.compute.manager [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Shelve offloading [ 1156.594954] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1156.595211] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4eae835-6855-4135-b7ea-db4770044e15 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.602361] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1156.602361] env[62368]: value = "task-1199256" [ 1156.602361] env[62368]: _type = "Task" [ 1156.602361] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.610193] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.687863] env[62368]: INFO nova.compute.manager [-] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Took 1.45 seconds to deallocate network for instance. [ 1156.744589] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.744913] env[62368]: DEBUG nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Instance network_info: |[{"id": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "address": "fa:16:3e:d6:e9:70", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a09cec0-45", "ovs_interfaceid": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1156.745364] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:e9:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a09cec0-45b4-46b9-872f-e153d6ea3791', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1156.753131] env[62368]: DEBUG oslo.service.loopingcall [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.753357] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1156.753581] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d743374a-df2e-4936-a81c-b0cb5e6be5dc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.773545] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1156.773545] env[62368]: value = "task-1199257" [ 1156.773545] env[62368]: _type = "Task" [ 1156.773545] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.781248] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199257, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.027663] env[62368]: DEBUG nova.compute.manager [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Received event network-changed-9a09cec0-45b4-46b9-872f-e153d6ea3791 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1157.028048] env[62368]: DEBUG nova.compute.manager [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Refreshing instance network info cache due to event network-changed-9a09cec0-45b4-46b9-872f-e153d6ea3791. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1157.028177] env[62368]: DEBUG oslo_concurrency.lockutils [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] Acquiring lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.028329] env[62368]: DEBUG oslo_concurrency.lockutils [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] Acquired lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.028491] env[62368]: DEBUG nova.network.neutron [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Refreshing network info cache for port 9a09cec0-45b4-46b9-872f-e153d6ea3791 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1157.051409] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e4d6dc-89a8-4395-b02e-791620e2380c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.059033] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb8a042-5534-475c-92e8-88661cfa8bd4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.091076] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3901de9-c7d3-4e02-aec6-820884f3fdfb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.099599] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e849fdc-c90e-4168-8872-7b431f27cfdc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.116967] env[62368]: DEBUG nova.compute.provider_tree [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.121852] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1157.122076] env[62368]: DEBUG nova.compute.manager [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1157.123075] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4392e4d1-a2c6-4631-8e83-cff08460f16e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.129454] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.129622] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.129804] env[62368]: DEBUG nova.network.neutron [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1157.194504] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.283216] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199257, 'name': CreateVM_Task, 'duration_secs': 0.327107} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.283348] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1157.284011] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.284197] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.284516] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1157.284756] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6825b47f-0c8f-40b5-82a2-9bb5cd7e3d48 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.289423] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1157.289423] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]527ca582-bf20-7882-66c1-9e65a4aeb188" [ 1157.289423] env[62368]: _type = "Task" [ 1157.289423] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.296573] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527ca582-bf20-7882-66c1-9e65a4aeb188, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.623944] env[62368]: DEBUG nova.scheduler.client.report [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1157.722639] env[62368]: DEBUG nova.network.neutron [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Updated VIF entry in instance network info cache for port 9a09cec0-45b4-46b9-872f-e153d6ea3791. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1157.722991] env[62368]: DEBUG nova.network.neutron [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Updating instance_info_cache with network_info: [{"id": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "address": "fa:16:3e:d6:e9:70", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a09cec0-45", "ovs_interfaceid": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.809248] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]527ca582-bf20-7882-66c1-9e65a4aeb188, 'name': SearchDatastore_Task, 'duration_secs': 0.010013} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.809248] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.809248] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1157.809248] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.809248] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.809248] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1157.809248] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-676506f6-2a5d-415c-bdbb-1a91b437f2ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.821018] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1157.821018] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1157.821018] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5159d4aa-b774-422b-9dfc-0944b5649b61 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.825237] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1157.825237] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52a77357-33fc-ede0-b923-6deaa8c660aa" [ 1157.825237] env[62368]: _type = "Task" [ 1157.825237] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.832560] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a77357-33fc-ede0-b923-6deaa8c660aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.879300] env[62368]: DEBUG nova.network.neutron [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [{"id": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "address": "fa:16:3e:0c:7d:cb", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8eb12d-c6", "ovs_interfaceid": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.128573] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.129577] env[62368]: DEBUG nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1158.131783] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.937s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.131974] env[62368]: DEBUG nova.objects.instance [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'resources' on Instance uuid 65e4f112-7cf3-4e21-b232-9c1c87ef1da3 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.226776] env[62368]: DEBUG oslo_concurrency.lockutils [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] Releasing lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.227104] env[62368]: DEBUG nova.compute.manager [req-26d57a63-d4da-490f-9bfc-ca7d33c6c4ee req-eb170125-92e3-463c-98dc-19d96b821b79 service nova] [instance: 65e4f112-7cf3-4e21-b232-9c1c87ef1da3] Received event network-vif-deleted-cb5acbed-41ed-4e72-b4f5-28aea7aa0d69 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1158.336372] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a77357-33fc-ede0-b923-6deaa8c660aa, 'name': SearchDatastore_Task, 'duration_secs': 0.013483} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.337137] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-543f8913-7357-4adc-aa0c-14f20933d304 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.342226] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1158.342226] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5226f758-98e0-f171-edfd-e553a2e76fbd" [ 1158.342226] env[62368]: _type = "Task" [ 1158.342226] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.349397] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5226f758-98e0-f171-edfd-e553a2e76fbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.381962] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.616696] env[62368]: DEBUG nova.compute.manager [req-639b52a7-3b18-4368-b9cc-b17fe8fe033e req-2c10cf96-9cc3-4188-a7d7-8793790729f0 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received event network-vif-unplugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1158.616955] env[62368]: DEBUG oslo_concurrency.lockutils [req-639b52a7-3b18-4368-b9cc-b17fe8fe033e req-2c10cf96-9cc3-4188-a7d7-8793790729f0 service nova] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.617186] env[62368]: DEBUG oslo_concurrency.lockutils [req-639b52a7-3b18-4368-b9cc-b17fe8fe033e req-2c10cf96-9cc3-4188-a7d7-8793790729f0 service nova] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.617361] env[62368]: DEBUG oslo_concurrency.lockutils [req-639b52a7-3b18-4368-b9cc-b17fe8fe033e req-2c10cf96-9cc3-4188-a7d7-8793790729f0 service nova] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.617531] env[62368]: DEBUG nova.compute.manager [req-639b52a7-3b18-4368-b9cc-b17fe8fe033e req-2c10cf96-9cc3-4188-a7d7-8793790729f0 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] No waiting events found dispatching network-vif-unplugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1158.617756] env[62368]: WARNING nova.compute.manager [req-639b52a7-3b18-4368-b9cc-b17fe8fe033e req-2c10cf96-9cc3-4188-a7d7-8793790729f0 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received unexpected event network-vif-unplugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 for instance with vm_state shelved and task_state shelving_offloading. [ 1158.634199] env[62368]: DEBUG nova.compute.utils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1158.635514] env[62368]: DEBUG nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1158.635740] env[62368]: DEBUG nova.network.neutron [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1158.682269] env[62368]: DEBUG nova.policy [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3ba554eb5934cfca29a3dc5a155ef2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0a61d319d764f52aff2e6f6c4723238', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1158.726764] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda7a84b-e23a-4972-8b65-f11504b9c8a9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.734570] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1158.735531] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5fbdaef-18b1-4199-b26e-735d218ca931 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.739263] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d4e036-ca88-4b16-9776-44c270fbee71 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.746122] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1158.771484] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-646c3dbd-b2ca-4f30-b070-d70398c21a01 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.773946] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3582670c-60a1-4216-b0f9-07f2daefdb17 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.780743] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d63d8a-4fc9-4224-a750-dbbb834bf71e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.793823] env[62368]: DEBUG nova.compute.provider_tree [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.840868] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1158.841103] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1158.841317] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleting the datastore file [datastore1] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.841544] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18e460ee-efd9-4c30-b6a6-334ac2583f73 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.852085] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5226f758-98e0-f171-edfd-e553a2e76fbd, 'name': SearchDatastore_Task, 'duration_secs': 0.010391} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.853189] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.853490] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606/9b792344-4397-4c8d-906e-64a4c7d91606.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1158.853823] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1158.853823] env[62368]: value = "task-1199259" [ 1158.853823] env[62368]: _type = "Task" [ 1158.853823] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.854087] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d600bb97-cebd-4a2a-a56e-d0a12359bfb1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.861932] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1158.861932] env[62368]: value = "task-1199260" [ 1158.861932] env[62368]: _type = "Task" [ 1158.861932] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.864738] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199259, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.872361] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.005821] env[62368]: DEBUG nova.network.neutron [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Successfully created port: eb8c3552-922e-4388-b342-afba75f4a3e0 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1159.139815] env[62368]: DEBUG nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1159.297437] env[62368]: DEBUG nova.scheduler.client.report [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1159.367542] env[62368]: DEBUG oslo_vmware.api [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199259, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152951} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.370567] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1159.370762] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1159.370938] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1159.377891] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50251} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.378146] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606/9b792344-4397-4c8d-906e-64a4c7d91606.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1159.378790] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1159.378790] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df4276ff-1bfa-45ad-a8bf-f461e6fee5f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.384332] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1159.384332] env[62368]: value = "task-1199261" [ 1159.384332] env[62368]: _type = "Task" [ 1159.384332] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.389607] env[62368]: INFO nova.scheduler.client.report [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted allocations for instance 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 [ 1159.395539] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199261, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.802675] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.822368] env[62368]: INFO nova.scheduler.client.report [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleted allocations for instance 65e4f112-7cf3-4e21-b232-9c1c87ef1da3 [ 1159.893346] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199261, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089371} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.893631] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1159.894426] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e710cf5b-eb7f-4298-b0cb-c6f1d7d16770 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.897655] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.898073] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.898321] env[62368]: DEBUG nova.objects.instance [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'resources' on Instance uuid 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.919654] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606/9b792344-4397-4c8d-906e-64a4c7d91606.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.920483] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34768103-e3f8-42b4-9d3a-971c4f40813c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.942085] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1159.942085] env[62368]: value = "task-1199262" [ 1159.942085] env[62368]: _type = "Task" [ 1159.942085] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.949793] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199262, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.150109] env[62368]: DEBUG nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1160.176090] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1160.176504] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1160.176778] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1160.177209] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1160.177483] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1160.177820] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1160.178198] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1160.178492] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1160.178805] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1160.179118] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1160.179434] env[62368]: DEBUG nova.virt.hardware [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1160.180744] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9d4c00-a50e-4c63-9091-f6688f35bfcf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.188665] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9712fd-563f-4f9b-9cee-870d533ad9e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.330537] env[62368]: DEBUG oslo_concurrency.lockutils [None req-c76e55a0-0379-4faa-8384-02dc01533988 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "65e4f112-7cf3-4e21-b232-9c1c87ef1da3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.225s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.402908] env[62368]: DEBUG nova.objects.instance [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'numa_topology' on Instance uuid 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.454080] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199262, 'name': ReconfigVM_Task, 'duration_secs': 0.311091} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.454080] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Reconfigured VM instance instance-00000072 to attach disk [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606/9b792344-4397-4c8d-906e-64a4c7d91606.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.454080] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92979532-5bfc-4213-b96e-4d1e5894c032 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.458923] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1160.458923] env[62368]: value = "task-1199263" [ 1160.458923] env[62368]: _type = "Task" [ 1160.458923] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.465237] env[62368]: DEBUG nova.compute.manager [req-4042a555-077a-4ba6-88ac-9d8354133673 req-0a47de56-b267-4135-aeb9-55b61e6050f1 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Received event network-vif-plugged-eb8c3552-922e-4388-b342-afba75f4a3e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1160.465445] env[62368]: DEBUG oslo_concurrency.lockutils [req-4042a555-077a-4ba6-88ac-9d8354133673 req-0a47de56-b267-4135-aeb9-55b61e6050f1 service nova] Acquiring lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.465666] env[62368]: DEBUG oslo_concurrency.lockutils [req-4042a555-077a-4ba6-88ac-9d8354133673 req-0a47de56-b267-4135-aeb9-55b61e6050f1 service nova] Lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.465870] env[62368]: DEBUG oslo_concurrency.lockutils [req-4042a555-077a-4ba6-88ac-9d8354133673 req-0a47de56-b267-4135-aeb9-55b61e6050f1 service nova] Lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.466104] env[62368]: DEBUG nova.compute.manager [req-4042a555-077a-4ba6-88ac-9d8354133673 req-0a47de56-b267-4135-aeb9-55b61e6050f1 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] No waiting events found dispatching network-vif-plugged-eb8c3552-922e-4388-b342-afba75f4a3e0 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1160.466496] env[62368]: WARNING nova.compute.manager [req-4042a555-077a-4ba6-88ac-9d8354133673 req-0a47de56-b267-4135-aeb9-55b61e6050f1 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Received unexpected event network-vif-plugged-eb8c3552-922e-4388-b342-afba75f4a3e0 for instance with vm_state building and task_state spawning. [ 1160.470090] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199263, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.632866] env[62368]: DEBUG nova.network.neutron [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Successfully updated port: eb8c3552-922e-4388-b342-afba75f4a3e0 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1160.655038] env[62368]: DEBUG nova.compute.manager [req-8767a40d-a0f7-4f00-b5db-4c51cd976f8f req-09e60517-f0a0-4043-9d19-653029ea9e3d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received event network-changed-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1160.655213] env[62368]: DEBUG nova.compute.manager [req-8767a40d-a0f7-4f00-b5db-4c51cd976f8f req-09e60517-f0a0-4043-9d19-653029ea9e3d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Refreshing instance network info cache due to event network-changed-7e8eb12d-c67a-4b59-9e66-b3b530e66e75. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1160.655428] env[62368]: DEBUG oslo_concurrency.lockutils [req-8767a40d-a0f7-4f00-b5db-4c51cd976f8f req-09e60517-f0a0-4043-9d19-653029ea9e3d service nova] Acquiring lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.655575] env[62368]: DEBUG oslo_concurrency.lockutils [req-8767a40d-a0f7-4f00-b5db-4c51cd976f8f req-09e60517-f0a0-4043-9d19-653029ea9e3d service nova] Acquired lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.655738] env[62368]: DEBUG nova.network.neutron [req-8767a40d-a0f7-4f00-b5db-4c51cd976f8f req-09e60517-f0a0-4043-9d19-653029ea9e3d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Refreshing network info cache for port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1160.904984] env[62368]: DEBUG nova.objects.base [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Object Instance<27ff9bed-7c5f-4553-9d8e-8ec78708caf1> lazy-loaded attributes: resources,numa_topology {{(pid=62368) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1160.969065] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199263, 'name': Rename_Task, 'duration_secs': 0.150576} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.970115] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1160.970833] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297475df-91ac-4491-9653-75dd37a69554 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.973181] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcbaa555-532d-4c36-ae77-806b74f09b60 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.978788] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883e7670-e48b-488b-a68c-65a52a38aec9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.983042] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1160.983042] env[62368]: value = "task-1199264" [ 1160.983042] env[62368]: _type = "Task" [ 1160.983042] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.010201] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b34f95-72c6-45f4-b5b3-ee9845daf49d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.016072] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199264, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.021245] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57f57cd-d171-40b4-9fba-acadc817432d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.034950] env[62368]: DEBUG nova.compute.provider_tree [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.136012] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquiring lock "refresh_cache-c58ed4ff-0013-401e-ab62-e1ba5291bfbd" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.136185] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquired lock "refresh_cache-c58ed4ff-0013-401e-ab62-e1ba5291bfbd" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.136324] env[62368]: DEBUG nova.network.neutron [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1161.294238] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b4070763-a2f2-480f-909c-e323f15a9ac5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.294537] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.389142] env[62368]: DEBUG nova.network.neutron [req-8767a40d-a0f7-4f00-b5db-4c51cd976f8f req-09e60517-f0a0-4043-9d19-653029ea9e3d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updated VIF entry in instance network info cache for port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1161.389520] env[62368]: DEBUG nova.network.neutron [req-8767a40d-a0f7-4f00-b5db-4c51cd976f8f req-09e60517-f0a0-4043-9d19-653029ea9e3d service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [{"id": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "address": "fa:16:3e:0c:7d:cb", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap7e8eb12d-c6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.492680] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199264, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.538298] env[62368]: DEBUG nova.scheduler.client.report [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1161.674823] env[62368]: DEBUG nova.network.neutron [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1161.797673] env[62368]: INFO nova.compute.manager [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Detaching volume cb82f49c-702d-4495-9424-daff96816034 [ 1161.800754] env[62368]: DEBUG nova.network.neutron [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Updating instance_info_cache with network_info: [{"id": "eb8c3552-922e-4388-b342-afba75f4a3e0", "address": "fa:16:3e:ae:e3:b8", "network": {"id": "a1d4299a-f5d2-4feb-874f-3d661ed23509", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1675858590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0a61d319d764f52aff2e6f6c4723238", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb8c3552-92", "ovs_interfaceid": "eb8c3552-922e-4388-b342-afba75f4a3e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.832926] env[62368]: INFO nova.virt.block_device [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Attempting to driver detach volume cb82f49c-702d-4495-9424-daff96816034 from mountpoint /dev/sdb [ 1161.833194] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1161.833386] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259911', 'volume_id': 'cb82f49c-702d-4495-9424-daff96816034', 'name': 'volume-cb82f49c-702d-4495-9424-daff96816034', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b4070763-a2f2-480f-909c-e323f15a9ac5', 'attached_at': '', 'detached_at': '', 'volume_id': 'cb82f49c-702d-4495-9424-daff96816034', 'serial': 'cb82f49c-702d-4495-9424-daff96816034'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1161.834329] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b5742e-0be1-4fb2-9204-885c98fdd22a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.856705] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f69915-663a-409c-98e4-c8adb043041a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.863365] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95e236c-e6aa-441e-8e2d-1d7b375b869c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.882399] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d57e05-1c80-442b-bde8-146ef8acfdf5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.898416] env[62368]: DEBUG oslo_concurrency.lockutils [req-8767a40d-a0f7-4f00-b5db-4c51cd976f8f req-09e60517-f0a0-4043-9d19-653029ea9e3d service nova] Releasing lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.898824] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] The volume has not been displaced from its original location: [datastore1] volume-cb82f49c-702d-4495-9424-daff96816034/volume-cb82f49c-702d-4495-9424-daff96816034.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1161.904030] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1161.904519] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e46b49d-722f-49c9-8a32-a956b07e68c4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.921362] env[62368]: DEBUG oslo_vmware.api [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1161.921362] env[62368]: value = "task-1199265" [ 1161.921362] env[62368]: _type = "Task" [ 1161.921362] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.928906] env[62368]: DEBUG oslo_vmware.api [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.992803] env[62368]: DEBUG oslo_vmware.api [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199264, 'name': PowerOnVM_Task, 'duration_secs': 0.742651} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.993070] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1161.993290] env[62368]: INFO nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Took 7.37 seconds to spawn the instance on the hypervisor. [ 1161.993473] env[62368]: DEBUG nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1161.994287] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff039b19-5467-4c72-a3fc-8778c8bde089 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.043523] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.145s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.146342] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.302589] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Releasing lock "refresh_cache-c58ed4ff-0013-401e-ab62-e1ba5291bfbd" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.302907] env[62368]: DEBUG nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Instance network_info: |[{"id": "eb8c3552-922e-4388-b342-afba75f4a3e0", "address": "fa:16:3e:ae:e3:b8", "network": {"id": "a1d4299a-f5d2-4feb-874f-3d661ed23509", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1675858590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0a61d319d764f52aff2e6f6c4723238", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb8c3552-92", "ovs_interfaceid": "eb8c3552-922e-4388-b342-afba75f4a3e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1162.303365] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:e3:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1cbd5e0e-9116-46f1-9748-13a73d2d7e75', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb8c3552-922e-4388-b342-afba75f4a3e0', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1162.311491] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Creating folder: Project (a0a61d319d764f52aff2e6f6c4723238). Parent ref: group-v259706. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1162.312054] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-64831d62-38b3-4315-90b9-75cec6fdf9a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.322590] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Created folder: Project (a0a61d319d764f52aff2e6f6c4723238) in parent group-v259706. [ 1162.322787] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Creating folder: Instances. Parent ref: group-v259924. {{(pid=62368) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1162.323038] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d946a1f8-453f-4a6b-a43d-ab18493ef868 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.331791] env[62368]: INFO nova.virt.vmwareapi.vm_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Created folder: Instances in parent group-v259924. [ 1162.332147] env[62368]: DEBUG oslo.service.loopingcall [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1162.332360] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1162.332575] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb47f857-252f-4569-b302-7ff27b60f023 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.350633] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1162.350633] env[62368]: value = "task-1199268" [ 1162.350633] env[62368]: _type = "Task" [ 1162.350633] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.357933] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199268, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.431428] env[62368]: DEBUG oslo_vmware.api [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199265, 'name': ReconfigVM_Task, 'duration_secs': 0.245364} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.431723] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1162.436503] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acdf5904-92a2-46ac-b2ff-5eb4a74a8064 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.452113] env[62368]: DEBUG oslo_vmware.api [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1162.452113] env[62368]: value = "task-1199269" [ 1162.452113] env[62368]: _type = "Task" [ 1162.452113] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.460142] env[62368]: DEBUG oslo_vmware.api [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199269, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.492253] env[62368]: DEBUG nova.compute.manager [req-f6d68c59-99a5-4d90-b1f0-01b443f762ce req-4cd9e096-6a75-4e4a-9714-c6f81b6ca359 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Received event network-changed-eb8c3552-922e-4388-b342-afba75f4a3e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1162.492466] env[62368]: DEBUG nova.compute.manager [req-f6d68c59-99a5-4d90-b1f0-01b443f762ce req-4cd9e096-6a75-4e4a-9714-c6f81b6ca359 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Refreshing instance network info cache due to event network-changed-eb8c3552-922e-4388-b342-afba75f4a3e0. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1162.492689] env[62368]: DEBUG oslo_concurrency.lockutils [req-f6d68c59-99a5-4d90-b1f0-01b443f762ce req-4cd9e096-6a75-4e4a-9714-c6f81b6ca359 service nova] Acquiring lock "refresh_cache-c58ed4ff-0013-401e-ab62-e1ba5291bfbd" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.492850] env[62368]: DEBUG oslo_concurrency.lockutils [req-f6d68c59-99a5-4d90-b1f0-01b443f762ce req-4cd9e096-6a75-4e4a-9714-c6f81b6ca359 service nova] Acquired lock "refresh_cache-c58ed4ff-0013-401e-ab62-e1ba5291bfbd" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.493055] env[62368]: DEBUG nova.network.neutron [req-f6d68c59-99a5-4d90-b1f0-01b443f762ce req-4cd9e096-6a75-4e4a-9714-c6f81b6ca359 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Refreshing network info cache for port eb8c3552-922e-4388-b342-afba75f4a3e0 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1162.511793] env[62368]: INFO nova.compute.manager [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Took 12.10 seconds to build instance. [ 1162.551041] env[62368]: DEBUG oslo_concurrency.lockutils [None req-28d1f8b4-500d-469f-af31-32366206a24d tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.597s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.552842] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.407s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.553151] env[62368]: INFO nova.compute.manager [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Unshelving [ 1162.784201] env[62368]: INFO nova.compute.manager [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Rescuing [ 1162.784408] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.784566] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.784746] env[62368]: DEBUG nova.network.neutron [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1162.860881] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199268, 'name': CreateVM_Task, 'duration_secs': 0.337506} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.860881] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1162.861582] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.861775] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.862147] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1162.862406] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-009e1a85-8cee-4f5a-a1d4-42eed3399f13 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.867612] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1162.867612] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]524cf2d1-f5fd-5a90-225f-68d3bb2ccf9f" [ 1162.867612] env[62368]: _type = "Task" [ 1162.867612] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.875047] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524cf2d1-f5fd-5a90-225f-68d3bb2ccf9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.961592] env[62368]: DEBUG oslo_vmware.api [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199269, 'name': ReconfigVM_Task, 'duration_secs': 0.153334} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.961902] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259911', 'volume_id': 'cb82f49c-702d-4495-9424-daff96816034', 'name': 'volume-cb82f49c-702d-4495-9424-daff96816034', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b4070763-a2f2-480f-909c-e323f15a9ac5', 'attached_at': '', 'detached_at': '', 'volume_id': 'cb82f49c-702d-4495-9424-daff96816034', 'serial': 'cb82f49c-702d-4495-9424-daff96816034'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1163.013783] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb278f78-a09b-4023-a7f5-4a243b1b75f4 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "9b792344-4397-4c8d-906e-64a4c7d91606" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.607s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.193772] env[62368]: DEBUG nova.network.neutron [req-f6d68c59-99a5-4d90-b1f0-01b443f762ce req-4cd9e096-6a75-4e4a-9714-c6f81b6ca359 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Updated VIF entry in instance network info cache for port eb8c3552-922e-4388-b342-afba75f4a3e0. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1163.194164] env[62368]: DEBUG nova.network.neutron [req-f6d68c59-99a5-4d90-b1f0-01b443f762ce req-4cd9e096-6a75-4e4a-9714-c6f81b6ca359 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Updating instance_info_cache with network_info: [{"id": "eb8c3552-922e-4388-b342-afba75f4a3e0", "address": "fa:16:3e:ae:e3:b8", "network": {"id": "a1d4299a-f5d2-4feb-874f-3d661ed23509", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1675858590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a0a61d319d764f52aff2e6f6c4723238", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb8c3552-92", "ovs_interfaceid": "eb8c3552-922e-4388-b342-afba75f4a3e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.378583] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]524cf2d1-f5fd-5a90-225f-68d3bb2ccf9f, 'name': SearchDatastore_Task, 'duration_secs': 0.011657} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.379115] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.379115] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1163.379316] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.379467] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.379685] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1163.379968] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c10fe6c-aa68-4f61-ac26-5dfc836f867c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.391256] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1163.391495] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1163.392253] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6c30444-5ffb-4ac0-b82c-10adbd28098d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.400585] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1163.400585] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52e3f0f9-b157-a012-5fe1-255edae7c923" [ 1163.400585] env[62368]: _type = "Task" [ 1163.400585] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.415638] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52e3f0f9-b157-a012-5fe1-255edae7c923, 'name': SearchDatastore_Task, 'duration_secs': 0.009118} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.416452] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a1f3c78-f34c-4087-9b9f-6b93754c797b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.421669] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1163.421669] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52a9f4c6-421f-3a8b-e4c6-501731db55f2" [ 1163.421669] env[62368]: _type = "Task" [ 1163.421669] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.431595] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a9f4c6-421f-3a8b-e4c6-501731db55f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.502012] env[62368]: DEBUG nova.objects.instance [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'flavor' on Instance uuid b4070763-a2f2-480f-909c-e323f15a9ac5 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.520569] env[62368]: DEBUG nova.network.neutron [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Updating instance_info_cache with network_info: [{"id": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "address": "fa:16:3e:d6:e9:70", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a09cec0-45", "ovs_interfaceid": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.576163] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.576438] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.576651] env[62368]: DEBUG nova.objects.instance [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'pci_requests' on Instance uuid 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.697023] env[62368]: DEBUG oslo_concurrency.lockutils [req-f6d68c59-99a5-4d90-b1f0-01b443f762ce req-4cd9e096-6a75-4e4a-9714-c6f81b6ca359 service nova] Releasing lock "refresh_cache-c58ed4ff-0013-401e-ab62-e1ba5291bfbd" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.932560] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52a9f4c6-421f-3a8b-e4c6-501731db55f2, 'name': SearchDatastore_Task, 'duration_secs': 0.008602} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.932793] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.933082] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] c58ed4ff-0013-401e-ab62-e1ba5291bfbd/c58ed4ff-0013-401e-ab62-e1ba5291bfbd.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1163.933361] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c801368f-6f5e-4187-b77e-108ec35f5f2e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.940245] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1163.940245] env[62368]: value = "task-1199270" [ 1163.940245] env[62368]: _type = "Task" [ 1163.940245] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.948585] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.023464] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.080686] env[62368]: DEBUG nova.objects.instance [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'numa_topology' on Instance uuid 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.449768] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199270, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478915} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.450200] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] c58ed4ff-0013-401e-ab62-e1ba5291bfbd/c58ed4ff-0013-401e-ab62-e1ba5291bfbd.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1164.450298] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1164.450597] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2fbc21e-1763-4276-8225-098fbcae9508 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.457568] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1164.457568] env[62368]: value = "task-1199271" [ 1164.457568] env[62368]: _type = "Task" [ 1164.457568] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.466310] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199271, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.511680] env[62368]: DEBUG oslo_concurrency.lockutils [None req-a98161e8-f43a-40ab-a708-09d329d2e45e tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.217s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.559605] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1164.559863] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9117f94-bf8d-4a4e-a803-f61fc1dc2fc9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.567288] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1164.567288] env[62368]: value = "task-1199272" [ 1164.567288] env[62368]: _type = "Task" [ 1164.567288] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.577296] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.583816] env[62368]: INFO nova.compute.claims [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1164.640400] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b4070763-a2f2-480f-909c-e323f15a9ac5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.640668] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.640891] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b4070763-a2f2-480f-909c-e323f15a9ac5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.641093] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.641275] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.643437] env[62368]: INFO nova.compute.manager [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Terminating instance [ 1164.645209] env[62368]: DEBUG nova.compute.manager [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1164.645409] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1164.646266] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f510e2f6-1d5c-46cb-9be9-369f7a159a73 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.654073] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1164.654316] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-437454d3-8b62-4c20-aece-975f6546b522 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.661195] env[62368]: DEBUG oslo_vmware.api [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1164.661195] env[62368]: value = "task-1199273" [ 1164.661195] env[62368]: _type = "Task" [ 1164.661195] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.669538] env[62368]: DEBUG oslo_vmware.api [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199273, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.967944] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199271, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072086} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.968343] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1164.969161] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf292fa-3d02-4ec8-a1e6-ff2be7eaa3c8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.991453] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] c58ed4ff-0013-401e-ab62-e1ba5291bfbd/c58ed4ff-0013-401e-ab62-e1ba5291bfbd.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1164.991831] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b0b0ab6-2821-49fa-a3cd-4786cfb0d7a6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.012508] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1165.012508] env[62368]: value = "task-1199274" [ 1165.012508] env[62368]: _type = "Task" [ 1165.012508] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.019958] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199274, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.076676] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199272, 'name': PowerOffVM_Task, 'duration_secs': 0.289053} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.076939] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1165.077690] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d804455-fd01-4715-bb7c-2ff88d1f2731 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.098070] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b3566e-7d08-4e8a-a4ed-ab6172ea1c97 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.125100] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1165.125590] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79277308-1fe7-476d-8576-fe8cebef8db0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.132030] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1165.132030] env[62368]: value = "task-1199275" [ 1165.132030] env[62368]: _type = "Task" [ 1165.132030] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.140056] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.169134] env[62368]: DEBUG oslo_vmware.api [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199273, 'name': PowerOffVM_Task, 'duration_secs': 0.26496} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.169395] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1165.169565] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1165.169790] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95171cc9-1a81-4a1b-8d81-2dec529be30d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.237543] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1165.237892] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1165.238144] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleting the datastore file [datastore2] b4070763-a2f2-480f-909c-e323f15a9ac5 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1165.238456] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-670476dc-39a6-43e8-b2f4-089d8d1433f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.245550] env[62368]: DEBUG oslo_vmware.api [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1165.245550] env[62368]: value = "task-1199277" [ 1165.245550] env[62368]: _type = "Task" [ 1165.245550] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.253717] env[62368]: DEBUG oslo_vmware.api [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199277, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.521495] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199274, 'name': ReconfigVM_Task, 'duration_secs': 0.333855} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.521847] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Reconfigured VM instance instance-00000073 to attach disk [datastore1] c58ed4ff-0013-401e-ab62-e1ba5291bfbd/c58ed4ff-0013-401e-ab62-e1ba5291bfbd.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.522528] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-236fd891-dd6b-4525-88a6-9834ab42a2f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.529082] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1165.529082] env[62368]: value = "task-1199278" [ 1165.529082] env[62368]: _type = "Task" [ 1165.529082] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.536757] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199278, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.644535] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] VM already powered off {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1165.644804] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1165.645100] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.645294] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.645513] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1165.648268] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af52bef9-329a-4b23-9a07-60bd53ecd61d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.663433] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1165.663656] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1165.664395] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff9686b0-785f-400a-9021-c22ea6c3d8b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.669502] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1165.669502] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52692173-ce53-af92-cd99-48771c27e72a" [ 1165.669502] env[62368]: _type = "Task" [ 1165.669502] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.680271] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52692173-ce53-af92-cd99-48771c27e72a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.685910] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0d62a1-918e-4ef0-b280-d646cb862f75 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.692370] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a74432-f69e-43d1-9eaf-d1b133d5893b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.723609] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0044998-30a5-4f01-b20f-8f7850239190 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.730661] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470dd6d0-b5b6-4685-b6b1-a0a35f788079 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.743805] env[62368]: DEBUG nova.compute.provider_tree [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.753572] env[62368]: DEBUG oslo_vmware.api [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199277, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161441} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.754421] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1165.754615] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1165.754792] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1165.754965] env[62368]: INFO nova.compute.manager [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1165.755215] env[62368]: DEBUG oslo.service.loopingcall [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1165.755402] env[62368]: DEBUG nova.compute.manager [-] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1165.755497] env[62368]: DEBUG nova.network.neutron [-] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1166.039666] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199278, 'name': Rename_Task, 'duration_secs': 0.14385} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.039965] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1166.040246] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3224ac16-2c94-45ad-8617-63997c0b6f16 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.047309] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1166.047309] env[62368]: value = "task-1199279" [ 1166.047309] env[62368]: _type = "Task" [ 1166.047309] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.054931] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199279, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.163342] env[62368]: DEBUG nova.compute.manager [req-e21187f4-4503-4eee-a378-29f96f7469b0 req-b8894fab-2e31-420b-ba9e-1d635baa0ed8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Received event network-vif-deleted-dbd133bd-2126-425e-9772-cbff9f5b8ff7 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1166.163475] env[62368]: INFO nova.compute.manager [req-e21187f4-4503-4eee-a378-29f96f7469b0 req-b8894fab-2e31-420b-ba9e-1d635baa0ed8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Neutron deleted interface dbd133bd-2126-425e-9772-cbff9f5b8ff7; detaching it from the instance and deleting it from the info cache [ 1166.163650] env[62368]: DEBUG nova.network.neutron [req-e21187f4-4503-4eee-a378-29f96f7469b0 req-b8894fab-2e31-420b-ba9e-1d635baa0ed8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.180381] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52692173-ce53-af92-cd99-48771c27e72a, 'name': SearchDatastore_Task, 'duration_secs': 0.03343} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.181419] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4604a50d-cf12-4e6c-abe4-b39e20b9046f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.186441] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1166.186441] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5233f6d6-f120-cb70-787c-95779a0a39ec" [ 1166.186441] env[62368]: _type = "Task" [ 1166.186441] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.194158] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5233f6d6-f120-cb70-787c-95779a0a39ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.246849] env[62368]: DEBUG nova.scheduler.client.report [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1166.557184] env[62368]: DEBUG oslo_vmware.api [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199279, 'name': PowerOnVM_Task, 'duration_secs': 0.445225} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.557479] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1166.557667] env[62368]: INFO nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Took 6.41 seconds to spawn the instance on the hypervisor. [ 1166.557927] env[62368]: DEBUG nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1166.558699] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e724fe-f50f-440c-a465-a739f6dbaa56 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.641049] env[62368]: DEBUG nova.network.neutron [-] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.666077] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a6cd184f-d0da-41e9-a3c1-31723001303d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.676624] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd322748-e479-4dd1-936f-7002e0c35764 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.696025] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5233f6d6-f120-cb70-787c-95779a0a39ec, 'name': SearchDatastore_Task, 'duration_secs': 0.009333} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.700066] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.700344] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. {{(pid=62368) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1166.700677] env[62368]: DEBUG nova.compute.manager [req-e21187f4-4503-4eee-a378-29f96f7469b0 req-b8894fab-2e31-420b-ba9e-1d635baa0ed8 service nova] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Detach interface failed, port_id=dbd133bd-2126-425e-9772-cbff9f5b8ff7, reason: Instance b4070763-a2f2-480f-909c-e323f15a9ac5 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1166.701029] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b882e6e-9c0f-44cc-bd5f-82f040e4a011 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.707516] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1166.707516] env[62368]: value = "task-1199280" [ 1166.707516] env[62368]: _type = "Task" [ 1166.707516] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.714587] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199280, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.752250] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.175s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.781879] env[62368]: INFO nova.network.neutron [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1167.075832] env[62368]: INFO nova.compute.manager [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Took 11.16 seconds to build instance. [ 1167.143736] env[62368]: INFO nova.compute.manager [-] [instance: b4070763-a2f2-480f-909c-e323f15a9ac5] Took 1.39 seconds to deallocate network for instance. [ 1167.217607] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199280, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.578583] env[62368]: DEBUG oslo_concurrency.lockutils [None req-566cabf1-e32f-46e4-85e5-bacddd85a78a tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.672s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.650905] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.651235] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.651471] env[62368]: DEBUG nova.objects.instance [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'resources' on Instance uuid b4070763-a2f2-480f-909c-e323f15a9ac5 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.718714] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199280, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517168} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.719038] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk. [ 1167.719799] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59127be2-76a1-48a5-bcf8-0127acdf0c79 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.744012] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1167.744255] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07b3c11f-91b7-4d6f-9ce3-58061adb6819 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.764009] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1167.764009] env[62368]: value = "task-1199281" [ 1167.764009] env[62368]: _type = "Task" [ 1167.764009] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.771788] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199281, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.229708] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5ff97d-e0d4-4996-8b51-240cb7ddd2d9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.237877] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7fb1e1-5e27-4569-93c5-5b4fee218e84 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.270472] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5effb070-2d60-4aa1-9ec2-4f15a7d418f0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.277336] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199281, 'name': ReconfigVM_Task, 'duration_secs': 0.336511} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.279322] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Reconfigured VM instance instance-00000072 to attach disk [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d-rescue.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1168.280158] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcb1a69-bff6-4a0d-9519-77fcd6da39ac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.283329] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fbbe1d-9825-4e68-8f3d-b74102abe4a2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.287129] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquiring lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.287365] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.287570] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquiring lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.287755] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.287966] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.290652] env[62368]: INFO nova.compute.manager [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Terminating instance [ 1168.319484] env[62368]: DEBUG nova.compute.provider_tree [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1168.320915] env[62368]: DEBUG nova.compute.manager [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1168.321135] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1168.321401] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3c0376f-b703-4ab8-9b6c-10b0c0d226f8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.331719] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f08c45-f4ae-4ca1-9fed-5d63dbbb156c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.335226] env[62368]: DEBUG nova.scheduler.client.report [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1168.344319] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1168.345509] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b895b74e-0323-4a63-ac29-7c1745ebb72d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.347541] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1168.347541] env[62368]: value = "task-1199282" [ 1168.347541] env[62368]: _type = "Task" [ 1168.347541] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.352503] env[62368]: DEBUG oslo_vmware.api [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1168.352503] env[62368]: value = "task-1199283" [ 1168.352503] env[62368]: _type = "Task" [ 1168.352503] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.358957] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199282, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.365212] env[62368]: DEBUG oslo_vmware.api [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199283, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.454048] env[62368]: DEBUG nova.compute.manager [req-1bbd6cd0-2db6-404c-9e43-5f0f36ffbe6c req-a4760bd7-15ef-450e-85bc-801bc8231899 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received event network-vif-plugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1168.454323] env[62368]: DEBUG oslo_concurrency.lockutils [req-1bbd6cd0-2db6-404c-9e43-5f0f36ffbe6c req-a4760bd7-15ef-450e-85bc-801bc8231899 service nova] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.454536] env[62368]: DEBUG oslo_concurrency.lockutils [req-1bbd6cd0-2db6-404c-9e43-5f0f36ffbe6c req-a4760bd7-15ef-450e-85bc-801bc8231899 service nova] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.454708] env[62368]: DEBUG oslo_concurrency.lockutils [req-1bbd6cd0-2db6-404c-9e43-5f0f36ffbe6c req-a4760bd7-15ef-450e-85bc-801bc8231899 service nova] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.454943] env[62368]: DEBUG nova.compute.manager [req-1bbd6cd0-2db6-404c-9e43-5f0f36ffbe6c req-a4760bd7-15ef-450e-85bc-801bc8231899 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] No waiting events found dispatching network-vif-plugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1168.455198] env[62368]: WARNING nova.compute.manager [req-1bbd6cd0-2db6-404c-9e43-5f0f36ffbe6c req-a4760bd7-15ef-450e-85bc-801bc8231899 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received unexpected event network-vif-plugged-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 for instance with vm_state shelved_offloaded and task_state spawning. [ 1168.538875] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.539169] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.539328] env[62368]: DEBUG nova.network.neutron [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1168.843252] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.192s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.859368] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199282, 'name': ReconfigVM_Task, 'duration_secs': 0.164608} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.859885] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1168.860837] env[62368]: INFO nova.scheduler.client.report [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleted allocations for instance b4070763-a2f2-480f-909c-e323f15a9ac5 [ 1168.861814] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45222379-dd3f-4ffc-bf0a-cce3bdda5287 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.868180] env[62368]: DEBUG oslo_vmware.api [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199283, 'name': PowerOffVM_Task, 'duration_secs': 0.175826} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.868915] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1168.869152] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1168.869388] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29475931-bdc4-486b-ab5c-4590080f950e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.873471] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1168.873471] env[62368]: value = "task-1199284" [ 1168.873471] env[62368]: _type = "Task" [ 1168.873471] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.880405] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.932070] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1168.932318] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1168.932505] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Deleting the datastore file [datastore1] c58ed4ff-0013-401e-ab62-e1ba5291bfbd {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1168.932773] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d77dc75-4975-4184-9f8b-bc5007c0f47b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.939324] env[62368]: DEBUG oslo_vmware.api [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for the task: (returnval){ [ 1168.939324] env[62368]: value = "task-1199286" [ 1168.939324] env[62368]: _type = "Task" [ 1168.939324] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.946963] env[62368]: DEBUG oslo_vmware.api [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199286, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.301943] env[62368]: DEBUG nova.network.neutron [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [{"id": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "address": "fa:16:3e:0c:7d:cb", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8eb12d-c6", "ovs_interfaceid": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.371644] env[62368]: DEBUG oslo_concurrency.lockutils [None req-aaeeeec5-cc1b-467d-a836-13926d3eef41 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b4070763-a2f2-480f-909c-e323f15a9ac5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.731s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.382471] env[62368]: DEBUG oslo_vmware.api [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199284, 'name': PowerOnVM_Task, 'duration_secs': 0.409616} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.383194] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1169.385693] env[62368]: DEBUG nova.compute.manager [None req-e7bf05f6-d36b-48d1-b9f6-eb7d8887ca30 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1169.386644] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08d35a4-8bb3-4065-9b1d-9e887cad57ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.449061] env[62368]: DEBUG oslo_vmware.api [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Task: {'id': task-1199286, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152757} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.449328] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1169.449519] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1169.449697] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1169.449871] env[62368]: INFO nova.compute.manager [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1169.450132] env[62368]: DEBUG oslo.service.loopingcall [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1169.450324] env[62368]: DEBUG nova.compute.manager [-] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1169.450420] env[62368]: DEBUG nova.network.neutron [-] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1169.804377] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.832855] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3c36e38882ad3a400566ea39e8f24247',container_format='bare',created_at=2024-10-08T23:23:22Z,direct_url=,disk_format='vmdk',id=cf2811ab-2887-48e1-8edf-05289097fce3,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-2044034859-shelved',owner='6b7091fbcb2f4879b78075ed06b95c0d',properties=ImageMetaProps,protected=,size=31666176,status='active',tags=,updated_at=2024-10-08T23:23:35Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1169.833131] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1169.833329] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1169.833482] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1169.833631] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1169.833781] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1169.833998] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1169.834190] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1169.834364] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1169.834534] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1169.834714] env[62368]: DEBUG nova.virt.hardware [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1169.835596] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fccea8c8-c910-4a77-aacc-19b9a997d0a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.843153] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db81a5d8-e4dd-4f82-ab99-09f38dd2e6a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.856069] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:7d:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e8eb12d-c67a-4b59-9e66-b3b530e66e75', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1169.863447] env[62368]: DEBUG oslo.service.loopingcall [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1169.863688] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1169.863888] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ebcd5407-d48b-4ee6-9bb5-842a393b3007 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.883887] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1169.883887] env[62368]: value = "task-1199287" [ 1169.883887] env[62368]: _type = "Task" [ 1169.883887] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.891392] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199287, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.157022] env[62368]: DEBUG nova.network.neutron [-] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.394476] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199287, 'name': CreateVM_Task, 'duration_secs': 0.305956} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.394653] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1170.395326] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.395500] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.395930] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1170.396196] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e647fe0b-4181-4eef-9ba8-222cd5e33b10 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.400639] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1170.400639] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52268b47-ecfe-518e-f449-b1a22781a558" [ 1170.400639] env[62368]: _type = "Task" [ 1170.400639] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.407977] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52268b47-ecfe-518e-f449-b1a22781a558, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.481344] env[62368]: DEBUG nova.compute.manager [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received event network-changed-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1170.481544] env[62368]: DEBUG nova.compute.manager [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Refreshing instance network info cache due to event network-changed-7e8eb12d-c67a-4b59-9e66-b3b530e66e75. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1170.481877] env[62368]: DEBUG oslo_concurrency.lockutils [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] Acquiring lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.482043] env[62368]: DEBUG oslo_concurrency.lockutils [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] Acquired lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.482215] env[62368]: DEBUG nova.network.neutron [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Refreshing network info cache for port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1170.661296] env[62368]: INFO nova.compute.manager [-] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Took 1.21 seconds to deallocate network for instance. [ 1170.819872] env[62368]: INFO nova.compute.manager [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Unrescuing [ 1170.820259] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.820418] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquired lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.820590] env[62368]: DEBUG nova.network.neutron [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1170.910754] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.911035] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Processing image cf2811ab-2887-48e1-8edf-05289097fce3 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1170.911276] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3/cf2811ab-2887-48e1-8edf-05289097fce3.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.911433] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3/cf2811ab-2887-48e1-8edf-05289097fce3.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.911613] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1170.911862] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-271997f9-02fa-43ef-beb8-30a498ff9165 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.919608] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1170.919787] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1170.920528] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82013abc-162d-4c51-95c4-91494732764f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.926722] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1170.926722] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5255072d-aee4-c095-b5bf-97e59ea9f2d4" [ 1170.926722] env[62368]: _type = "Task" [ 1170.926722] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.933656] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5255072d-aee4-c095-b5bf-97e59ea9f2d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.171210] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.171521] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.172406] env[62368]: DEBUG nova.objects.instance [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lazy-loading 'resources' on Instance uuid c58ed4ff-0013-401e-ab62-e1ba5291bfbd {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.204921] env[62368]: DEBUG nova.network.neutron [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updated VIF entry in instance network info cache for port 7e8eb12d-c67a-4b59-9e66-b3b530e66e75. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1171.205328] env[62368]: DEBUG nova.network.neutron [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [{"id": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "address": "fa:16:3e:0c:7d:cb", "network": {"id": "8e0d4fb9-5020-4d8a-b731-345b013e75af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-806950485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b7091fbcb2f4879b78075ed06b95c0d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8eb12d-c6", "ovs_interfaceid": "7e8eb12d-c67a-4b59-9e66-b3b530e66e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.440951] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Preparing fetch location {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1171.441321] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Fetch image to [datastore2] OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645/OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645.vmdk {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1171.441535] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Downloading stream optimized image cf2811ab-2887-48e1-8edf-05289097fce3 to [datastore2] OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645/OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645.vmdk on the data store datastore2 as vApp {{(pid=62368) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1171.441793] env[62368]: DEBUG nova.virt.vmwareapi.images [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Downloading image file data cf2811ab-2887-48e1-8edf-05289097fce3 to the ESX as VM named 'OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645' {{(pid=62368) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1171.530442] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1171.530442] env[62368]: value = "resgroup-9" [ 1171.530442] env[62368]: _type = "ResourcePool" [ 1171.530442] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1171.530813] env[62368]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9784e286-7f3d-4fc8-b805-cdedd1672524 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.558866] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lease: (returnval){ [ 1171.558866] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529a88e8-6005-ef62-1214-ecf5e08993b9" [ 1171.558866] env[62368]: _type = "HttpNfcLease" [ 1171.558866] env[62368]: } obtained for vApp import into resource pool (val){ [ 1171.558866] env[62368]: value = "resgroup-9" [ 1171.558866] env[62368]: _type = "ResourcePool" [ 1171.558866] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1171.559180] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the lease: (returnval){ [ 1171.559180] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529a88e8-6005-ef62-1214-ecf5e08993b9" [ 1171.559180] env[62368]: _type = "HttpNfcLease" [ 1171.559180] env[62368]: } to be ready. {{(pid=62368) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1171.565295] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1171.565295] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529a88e8-6005-ef62-1214-ecf5e08993b9" [ 1171.565295] env[62368]: _type = "HttpNfcLease" [ 1171.565295] env[62368]: } is initializing. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1171.570544] env[62368]: DEBUG nova.network.neutron [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Updating instance_info_cache with network_info: [{"id": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "address": "fa:16:3e:d6:e9:70", "network": {"id": "68369b40-ae6a-4863-8e9b-f57ec9a5da8c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1948146926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "09d9046c25044b48830356a82644d1b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a09cec0-45", "ovs_interfaceid": "9a09cec0-45b4-46b9-872f-e153d6ea3791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.700479] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b84d58f8-de40-493f-9586-705ebdc4b759" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.700745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.707514] env[62368]: DEBUG oslo_concurrency.lockutils [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] Releasing lock "refresh_cache-27ff9bed-7c5f-4553-9d8e-8ec78708caf1" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.707772] env[62368]: DEBUG nova.compute.manager [req-6d44b4f0-714e-43d7-8ef7-1a00f858e0f3 req-2c64d53f-36a1-4798-a527-92bb80afa701 service nova] [instance: c58ed4ff-0013-401e-ab62-e1ba5291bfbd] Received event network-vif-deleted-eb8c3552-922e-4388-b342-afba75f4a3e0 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1171.757160] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf72e466-5c53-4122-a5a5-0b39440dea4a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.764365] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b8490c-e8b0-42f8-90cf-5798dad10ba0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.797278] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15878489-ec0b-4551-85c7-21c3e9ffb2ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.804364] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c13672-3e87-4835-b249-30364b63bd74 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.817234] env[62368]: DEBUG nova.compute.provider_tree [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.067628] env[62368]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1172.067628] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529a88e8-6005-ef62-1214-ecf5e08993b9" [ 1172.067628] env[62368]: _type = "HttpNfcLease" [ 1172.067628] env[62368]: } is ready. {{(pid=62368) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1172.068138] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1172.068138] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]529a88e8-6005-ef62-1214-ecf5e08993b9" [ 1172.068138] env[62368]: _type = "HttpNfcLease" [ 1172.068138] env[62368]: }. {{(pid=62368) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1172.068663] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c7786b-a3e6-4d8a-b7d1-b961bd0c4236 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.072743] env[62368]: DEBUG oslo_concurrency.lockutils [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Releasing lock "refresh_cache-9b792344-4397-4c8d-906e-64a4c7d91606" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.073432] env[62368]: DEBUG nova.objects.instance [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lazy-loading 'flavor' on Instance uuid 9b792344-4397-4c8d-906e-64a4c7d91606 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.078590] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea72a0-5117-1690-d255-9fbb6a5fc00c/disk-0.vmdk from lease info. {{(pid=62368) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1172.078797] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating HTTP connection to write to file with size = 31666176 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea72a0-5117-1690-d255-9fbb6a5fc00c/disk-0.vmdk. {{(pid=62368) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1172.144882] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6595c63c-ca8d-4860-9f26-8c3005645f19 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.203413] env[62368]: DEBUG nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1172.320698] env[62368]: DEBUG nova.scheduler.client.report [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1172.586155] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4554a361-7f71-42e7-b3c1-3fdb58fc72a0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.610334] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1172.616643] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c889bdb3-7757-41a5-9288-3b6e3b6011e9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.623387] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1172.623387] env[62368]: value = "task-1199289" [ 1172.623387] env[62368]: _type = "Task" [ 1172.623387] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.633091] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199289, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.729350] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.826685] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.655s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.829227] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.100s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.830875] env[62368]: INFO nova.compute.claims [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1172.848755] env[62368]: INFO nova.scheduler.client.report [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Deleted allocations for instance c58ed4ff-0013-401e-ab62-e1ba5291bfbd [ 1173.138537] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199289, 'name': PowerOffVM_Task, 'duration_secs': 0.207603} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.141662] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1173.147075] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Reconfiguring VM instance instance-00000072 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1173.148627] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f5c9925-de50-4572-bf72-2de1f7c77346 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.171040] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1173.171040] env[62368]: value = "task-1199290" [ 1173.171040] env[62368]: _type = "Task" [ 1173.171040] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.183537] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199290, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.354632] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Completed reading data from the image iterator. {{(pid=62368) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1173.354919] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea72a0-5117-1690-d255-9fbb6a5fc00c/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1173.358429] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a12057-4be4-4e34-8f89-2cdc599d3a45 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.362199] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4343ae1e-a14d-40b3-8175-cc9ba47e5a95 tempest-ServerAddressesTestJSON-2111549012 tempest-ServerAddressesTestJSON-2111549012-project-member] Lock "c58ed4ff-0013-401e-ab62-e1ba5291bfbd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.075s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.367520] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea72a0-5117-1690-d255-9fbb6a5fc00c/disk-0.vmdk is in state: ready. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1173.367698] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea72a0-5117-1690-d255-9fbb6a5fc00c/disk-0.vmdk. {{(pid=62368) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1173.368035] env[62368]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-98588f78-ad38-4410-80ad-ac9f956217b5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.597059] env[62368]: DEBUG oslo_vmware.rw_handles [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea72a0-5117-1690-d255-9fbb6a5fc00c/disk-0.vmdk. {{(pid=62368) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1173.597283] env[62368]: INFO nova.virt.vmwareapi.images [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Downloaded image file data cf2811ab-2887-48e1-8edf-05289097fce3 [ 1173.598205] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf62d79e-0497-44dd-bffe-90cc9047c4b0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.613819] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65752e67-4a5e-40e1-934f-bb841804da3c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.636411] env[62368]: INFO nova.virt.vmwareapi.images [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] The imported VM was unregistered [ 1173.639026] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Caching image {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1173.639269] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Creating directory with path [datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1173.639538] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6fa3a3d9-5d80-48d0-8929-cb35a1b5e76a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.649655] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Created directory with path [datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3 {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1173.649851] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645/OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645.vmdk to [datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3/cf2811ab-2887-48e1-8edf-05289097fce3.vmdk. {{(pid=62368) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1173.650136] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-3e77454f-aa76-44b8-990a-b5f2903df978 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.656639] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1173.656639] env[62368]: value = "task-1199292" [ 1173.656639] env[62368]: _type = "Task" [ 1173.656639] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.664082] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199292, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.680895] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199290, 'name': ReconfigVM_Task, 'duration_secs': 0.234451} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.681400] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Reconfigured VM instance instance-00000072 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1173.681400] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1173.681625] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0849d581-fb3d-41d4-8358-7431c94d9657 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.686877] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1173.686877] env[62368]: value = "task-1199293" [ 1173.686877] env[62368]: _type = "Task" [ 1173.686877] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.700541] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199293, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.914043] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165f9c31-8e6d-44a3-b915-039b0ba5bff2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.922587] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd4cd04-b49b-4e77-b12d-99e2775f5174 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.955574] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5941632-b980-45e1-9007-9b6dd4aba620 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.964570] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00c7202-5b46-4640-bbaf-5b8e9b72ea95 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.980818] env[62368]: DEBUG nova.compute.provider_tree [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.167377] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199292, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.196450] env[62368]: DEBUG oslo_vmware.api [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199293, 'name': PowerOnVM_Task, 'duration_secs': 0.40173} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.196735] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1174.196972] env[62368]: DEBUG nova.compute.manager [None req-7c7d32b5-b30d-46a0-8118-d03d861db99e tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1174.197811] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2411da-d97f-4f11-9bcc-3276256636d4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.484986] env[62368]: DEBUG nova.scheduler.client.report [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1174.667918] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199292, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.990301] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.161s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.990924] env[62368]: DEBUG nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1175.169555] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199292, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.496944] env[62368]: DEBUG nova.compute.utils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1175.498612] env[62368]: DEBUG nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1175.498799] env[62368]: DEBUG nova.network.neutron [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1175.549500] env[62368]: DEBUG nova.policy [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac2df24b03d147f5a33d67a97bbb5bb2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e41fb7a4acf043df959806368ae94ef9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1175.586773] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "9b792344-4397-4c8d-906e-64a4c7d91606" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.587090] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "9b792344-4397-4c8d-906e-64a4c7d91606" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.587317] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "9b792344-4397-4c8d-906e-64a4c7d91606-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.587516] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "9b792344-4397-4c8d-906e-64a4c7d91606-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.587694] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "9b792344-4397-4c8d-906e-64a4c7d91606-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.590109] env[62368]: INFO nova.compute.manager [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Terminating instance [ 1175.592083] env[62368]: DEBUG nova.compute.manager [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1175.592291] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1175.593187] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4fda18-847e-4cf3-89de-b03e35c13af4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.602870] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1175.604965] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-978ed3fc-537f-4f50-afa6-e74e1dafc239 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.612398] env[62368]: DEBUG oslo_vmware.api [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1175.612398] env[62368]: value = "task-1199294" [ 1175.612398] env[62368]: _type = "Task" [ 1175.612398] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.620878] env[62368]: DEBUG oslo_vmware.api [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.669196] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199292, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.896716] env[62368]: DEBUG nova.network.neutron [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Successfully created port: 7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1176.005584] env[62368]: DEBUG nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1176.125315] env[62368]: DEBUG oslo_vmware.api [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199294, 'name': PowerOffVM_Task, 'duration_secs': 0.228997} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.125315] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1176.125315] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1176.125315] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9967abea-2d69-41fa-8df0-145fb927ede2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.169788] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199292, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.2526} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.172930] env[62368]: INFO nova.virt.vmwareapi.ds_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645/OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645.vmdk to [datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3/cf2811ab-2887-48e1-8edf-05289097fce3.vmdk. [ 1176.172930] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Cleaning up location [datastore2] OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645 {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1176.172930] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_aa40fb7e-a919-450f-8e07-5d4cd2ba0645 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1176.172930] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-737fb81e-c22b-4f33-929f-3d4fe7610876 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.180018] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1176.180018] env[62368]: value = "task-1199296" [ 1176.180018] env[62368]: _type = "Task" [ 1176.180018] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.185855] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199296, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.212708] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1176.212857] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1176.213082] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Deleting the datastore file [datastore1] 9b792344-4397-4c8d-906e-64a4c7d91606 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1176.213378] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f697eeb-d275-4d17-9ca2-f5266f01bb26 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.220019] env[62368]: DEBUG oslo_vmware.api [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1176.220019] env[62368]: value = "task-1199297" [ 1176.220019] env[62368]: _type = "Task" [ 1176.220019] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.227808] env[62368]: DEBUG oslo_vmware.api [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199297, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.687520] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199296, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035079} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.687707] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1176.687885] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3/cf2811ab-2887-48e1-8edf-05289097fce3.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.688189] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3/cf2811ab-2887-48e1-8edf-05289097fce3.vmdk to [datastore2] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1/27ff9bed-7c5f-4553-9d8e-8ec78708caf1.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1176.688449] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6194c976-17b7-408b-92c9-dc0850315df7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.695383] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1176.695383] env[62368]: value = "task-1199298" [ 1176.695383] env[62368]: _type = "Task" [ 1176.695383] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.703102] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199298, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.727926] env[62368]: DEBUG oslo_vmware.api [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199297, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141217} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.728187] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1176.728376] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1176.728555] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1176.728729] env[62368]: INFO nova.compute.manager [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1176.728994] env[62368]: DEBUG oslo.service.loopingcall [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1176.729213] env[62368]: DEBUG nova.compute.manager [-] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1176.729311] env[62368]: DEBUG nova.network.neutron [-] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1176.968040] env[62368]: DEBUG nova.compute.manager [req-760f3a07-4d44-4a94-a4c1-2eec45c2910c req-6661b6e1-98ce-4889-97fb-7be410e73e50 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Received event network-vif-deleted-9a09cec0-45b4-46b9-872f-e153d6ea3791 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1176.968311] env[62368]: INFO nova.compute.manager [req-760f3a07-4d44-4a94-a4c1-2eec45c2910c req-6661b6e1-98ce-4889-97fb-7be410e73e50 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Neutron deleted interface 9a09cec0-45b4-46b9-872f-e153d6ea3791; detaching it from the instance and deleting it from the info cache [ 1176.968502] env[62368]: DEBUG nova.network.neutron [req-760f3a07-4d44-4a94-a4c1-2eec45c2910c req-6661b6e1-98ce-4889-97fb-7be410e73e50 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.012883] env[62368]: DEBUG nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1177.041651] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1177.041888] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1177.042064] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1177.042277] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1177.042430] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1177.042578] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1177.042789] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1177.042952] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1177.043143] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1177.043315] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1177.043498] env[62368]: DEBUG nova.virt.hardware [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1177.044458] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95312e1c-425d-4d2d-9548-48490823be99 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.054287] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e80e099-3e66-4192-9331-971a3c1ae8ee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.204915] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199298, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.421428] env[62368]: DEBUG nova.compute.manager [req-2397c57a-78a6-4a39-bb24-73cbd4bf26ba req-4a84cf8e-a326-47c4-bf5a-af765940052a service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Received event network-vif-plugged-7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1177.421663] env[62368]: DEBUG oslo_concurrency.lockutils [req-2397c57a-78a6-4a39-bb24-73cbd4bf26ba req-4a84cf8e-a326-47c4-bf5a-af765940052a service nova] Acquiring lock "b84d58f8-de40-493f-9586-705ebdc4b759-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.421877] env[62368]: DEBUG oslo_concurrency.lockutils [req-2397c57a-78a6-4a39-bb24-73cbd4bf26ba req-4a84cf8e-a326-47c4-bf5a-af765940052a service nova] Lock "b84d58f8-de40-493f-9586-705ebdc4b759-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.422065] env[62368]: DEBUG oslo_concurrency.lockutils [req-2397c57a-78a6-4a39-bb24-73cbd4bf26ba req-4a84cf8e-a326-47c4-bf5a-af765940052a service nova] Lock "b84d58f8-de40-493f-9586-705ebdc4b759-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.422245] env[62368]: DEBUG nova.compute.manager [req-2397c57a-78a6-4a39-bb24-73cbd4bf26ba req-4a84cf8e-a326-47c4-bf5a-af765940052a service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] No waiting events found dispatching network-vif-plugged-7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1177.423014] env[62368]: WARNING nova.compute.manager [req-2397c57a-78a6-4a39-bb24-73cbd4bf26ba req-4a84cf8e-a326-47c4-bf5a-af765940052a service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Received unexpected event network-vif-plugged-7fc54e0c-c523-43cc-b1a5-4a35cdf98767 for instance with vm_state building and task_state spawning. [ 1177.443885] env[62368]: DEBUG nova.network.neutron [-] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.472492] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98302ce1-37fe-42e3-a001-c92b0a2f38a1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.482198] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a0ac7a-045f-484a-b0dd-2dff20c7f46a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.507625] env[62368]: DEBUG nova.compute.manager [req-760f3a07-4d44-4a94-a4c1-2eec45c2910c req-6661b6e1-98ce-4889-97fb-7be410e73e50 service nova] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Detach interface failed, port_id=9a09cec0-45b4-46b9-872f-e153d6ea3791, reason: Instance 9b792344-4397-4c8d-906e-64a4c7d91606 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1177.535279] env[62368]: DEBUG nova.network.neutron [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Successfully updated port: 7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1177.705823] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199298, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.946830] env[62368]: INFO nova.compute.manager [-] [instance: 9b792344-4397-4c8d-906e-64a4c7d91606] Took 1.22 seconds to deallocate network for instance. [ 1178.041683] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.041816] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.041959] env[62368]: DEBUG nova.network.neutron [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1178.207479] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199298, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.454618] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.455993] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.455993] env[62368]: DEBUG nova.objects.instance [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lazy-loading 'resources' on Instance uuid 9b792344-4397-4c8d-906e-64a4c7d91606 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1178.578932] env[62368]: DEBUG nova.network.neutron [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1178.711359] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199298, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.778486] env[62368]: DEBUG nova.network.neutron [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Updating instance_info_cache with network_info: [{"id": "7fc54e0c-c523-43cc-b1a5-4a35cdf98767", "address": "fa:16:3e:81:d9:f6", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fc54e0c-c5", "ovs_interfaceid": "7fc54e0c-c523-43cc-b1a5-4a35cdf98767", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.020252] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6383a51-997b-4a39-bafe-57c1247ae13e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.027308] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf5ca2b-270a-4ee9-8df6-772fdb802f74 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.057127] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db680206-fd51-47d0-ad0f-0d8b1a77d342 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.063952] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8337402a-a828-465b-80f0-18f9ceacbb27 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.076698] env[62368]: DEBUG nova.compute.provider_tree [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1179.209816] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199298, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.24465} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.210238] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cf2811ab-2887-48e1-8edf-05289097fce3/cf2811ab-2887-48e1-8edf-05289097fce3.vmdk to [datastore2] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1/27ff9bed-7c5f-4553-9d8e-8ec78708caf1.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1179.210922] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db15ae86-235c-49f8-aa3d-f54ed31b79be {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.232520] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1/27ff9bed-7c5f-4553-9d8e-8ec78708caf1.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1179.232779] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e73c95a-1f48-4008-933e-6df8b6d50acc {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.251985] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1179.251985] env[62368]: value = "task-1199299" [ 1179.251985] env[62368]: _type = "Task" [ 1179.251985] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.261538] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199299, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.281257] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.281532] env[62368]: DEBUG nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Instance network_info: |[{"id": "7fc54e0c-c523-43cc-b1a5-4a35cdf98767", "address": "fa:16:3e:81:d9:f6", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fc54e0c-c5", "ovs_interfaceid": "7fc54e0c-c523-43cc-b1a5-4a35cdf98767", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1179.281977] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:d9:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7fc54e0c-c523-43cc-b1a5-4a35cdf98767', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1179.289612] env[62368]: DEBUG oslo.service.loopingcall [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1179.289826] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1179.290081] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f277e639-1605-4800-8ae4-dd38455e647c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.308661] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1179.308661] env[62368]: value = "task-1199300" [ 1179.308661] env[62368]: _type = "Task" [ 1179.308661] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.315887] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199300, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.444830] env[62368]: DEBUG nova.compute.manager [req-988e8200-5000-41d0-8b14-780caa585f16 req-6fd4e211-dce6-4e1f-90dc-134fd75038ac service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Received event network-changed-7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1179.445084] env[62368]: DEBUG nova.compute.manager [req-988e8200-5000-41d0-8b14-780caa585f16 req-6fd4e211-dce6-4e1f-90dc-134fd75038ac service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Refreshing instance network info cache due to event network-changed-7fc54e0c-c523-43cc-b1a5-4a35cdf98767. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1179.445326] env[62368]: DEBUG oslo_concurrency.lockutils [req-988e8200-5000-41d0-8b14-780caa585f16 req-6fd4e211-dce6-4e1f-90dc-134fd75038ac service nova] Acquiring lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.445491] env[62368]: DEBUG oslo_concurrency.lockutils [req-988e8200-5000-41d0-8b14-780caa585f16 req-6fd4e211-dce6-4e1f-90dc-134fd75038ac service nova] Acquired lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.445676] env[62368]: DEBUG nova.network.neutron [req-988e8200-5000-41d0-8b14-780caa585f16 req-6fd4e211-dce6-4e1f-90dc-134fd75038ac service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Refreshing network info cache for port 7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1179.580652] env[62368]: DEBUG nova.scheduler.client.report [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1179.762138] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199299, 'name': ReconfigVM_Task, 'duration_secs': 0.292373} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.762981] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1/27ff9bed-7c5f-4553-9d8e-8ec78708caf1.vmdk or device None with type streamOptimized {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1179.763315] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15ba3088-e954-440d-bc9f-fda42ee4b109 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.769300] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1179.769300] env[62368]: value = "task-1199301" [ 1179.769300] env[62368]: _type = "Task" [ 1179.769300] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.776497] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199301, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.818850] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199300, 'name': CreateVM_Task, 'duration_secs': 0.364653} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.818850] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1179.818850] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.818850] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.819192] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1179.819434] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47599341-1506-49fd-851c-8b7ec23fc2ca {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.823584] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1179.823584] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5271c32f-530e-ea7a-d070-8e04e817af12" [ 1179.823584] env[62368]: _type = "Task" [ 1179.823584] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.830630] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5271c32f-530e-ea7a-d070-8e04e817af12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.085825] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.103496] env[62368]: INFO nova.scheduler.client.report [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Deleted allocations for instance 9b792344-4397-4c8d-906e-64a4c7d91606 [ 1180.127504] env[62368]: DEBUG nova.network.neutron [req-988e8200-5000-41d0-8b14-780caa585f16 req-6fd4e211-dce6-4e1f-90dc-134fd75038ac service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Updated VIF entry in instance network info cache for port 7fc54e0c-c523-43cc-b1a5-4a35cdf98767. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1180.127865] env[62368]: DEBUG nova.network.neutron [req-988e8200-5000-41d0-8b14-780caa585f16 req-6fd4e211-dce6-4e1f-90dc-134fd75038ac service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Updating instance_info_cache with network_info: [{"id": "7fc54e0c-c523-43cc-b1a5-4a35cdf98767", "address": "fa:16:3e:81:d9:f6", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fc54e0c-c5", "ovs_interfaceid": "7fc54e0c-c523-43cc-b1a5-4a35cdf98767", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.279215] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199301, 'name': Rename_Task, 'duration_secs': 0.158855} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.279548] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1180.279716] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-928fd441-e308-4c4e-ada4-be48942ec122 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.286510] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1180.286510] env[62368]: value = "task-1199302" [ 1180.286510] env[62368]: _type = "Task" [ 1180.286510] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.293581] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199302, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.335334] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5271c32f-530e-ea7a-d070-8e04e817af12, 'name': SearchDatastore_Task, 'duration_secs': 0.010605} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.335626] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.335861] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1180.336109] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.336264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.336446] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1180.336691] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53ddee1c-4ab4-43bd-bbb2-4e1ba87b2808 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.343681] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1180.343858] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1180.344535] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9fe222d-5e92-47c9-b48b-157ec0eec9e3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.349252] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1180.349252] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52cd9581-9de1-4fa7-8af4-bfed93e19a68" [ 1180.349252] env[62368]: _type = "Task" [ 1180.349252] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.356430] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cd9581-9de1-4fa7-8af4-bfed93e19a68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.613124] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e0c9b4d1-8e8f-48d8-a012-95c7d8cf9c0d tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "9b792344-4397-4c8d-906e-64a4c7d91606" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.026s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.630895] env[62368]: DEBUG oslo_concurrency.lockutils [req-988e8200-5000-41d0-8b14-780caa585f16 req-6fd4e211-dce6-4e1f-90dc-134fd75038ac service nova] Releasing lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.796987] env[62368]: DEBUG oslo_vmware.api [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199302, 'name': PowerOnVM_Task, 'duration_secs': 0.449753} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.797292] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1180.859709] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52cd9581-9de1-4fa7-8af4-bfed93e19a68, 'name': SearchDatastore_Task, 'duration_secs': 0.007752} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.860512] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a78dfd98-26eb-4ff3-a11b-a6d24cd2d424 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.865371] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1180.865371] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]5242ee7b-7c37-88c1-96f3-87bacd83a4a5" [ 1180.865371] env[62368]: _type = "Task" [ 1180.865371] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.872710] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5242ee7b-7c37-88c1-96f3-87bacd83a4a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.903985] env[62368]: DEBUG nova.compute.manager [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1180.905177] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1380cc26-2f73-4170-a291-7889b5f45842 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.145142] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.145408] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.145625] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.145829] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.146062] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.148206] env[62368]: INFO nova.compute.manager [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Terminating instance [ 1181.150048] env[62368]: DEBUG nova.compute.manager [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1181.150258] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1181.151070] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1542dd44-1e38-42f4-9fc4-93d4e443e5c8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.158744] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1181.158969] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04abc7db-3fd6-480c-a828-0f5f99e1e3b8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.164766] env[62368]: DEBUG oslo_vmware.api [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1181.164766] env[62368]: value = "task-1199303" [ 1181.164766] env[62368]: _type = "Task" [ 1181.164766] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.171780] env[62368]: DEBUG oslo_vmware.api [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.376174] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]5242ee7b-7c37-88c1-96f3-87bacd83a4a5, 'name': SearchDatastore_Task, 'duration_secs': 0.008701} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.376537] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.376674] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b84d58f8-de40-493f-9586-705ebdc4b759/b84d58f8-de40-493f-9586-705ebdc4b759.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1181.376898] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a7fca83-59e7-4a1f-b84c-918bb5908b77 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.382645] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1181.382645] env[62368]: value = "task-1199304" [ 1181.382645] env[62368]: _type = "Task" [ 1181.382645] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.390049] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199304, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.420024] env[62368]: DEBUG oslo_concurrency.lockutils [None req-03491f83-ae8d-4299-9e39-4110acbf4205 tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.867s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.676210] env[62368]: DEBUG oslo_vmware.api [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199303, 'name': PowerOffVM_Task, 'duration_secs': 0.222893} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.676527] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1181.676704] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1181.676973] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03413484-982e-49e2-8025-9c25f1646f29 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.757032] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1181.757259] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1181.757450] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Deleting the datastore file [datastore2] 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1181.757730] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9641ee33-d8ff-475d-9896-932f8ff212f4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.766506] env[62368]: DEBUG oslo_vmware.api [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for the task: (returnval){ [ 1181.766506] env[62368]: value = "task-1199306" [ 1181.766506] env[62368]: _type = "Task" [ 1181.766506] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.775088] env[62368]: DEBUG oslo_vmware.api [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.892422] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199304, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442328} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.892719] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore1] b84d58f8-de40-493f-9586-705ebdc4b759/b84d58f8-de40-493f-9586-705ebdc4b759.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1181.892938] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1181.893233] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29f00ae0-803d-4f38-9645-617db64d6b75 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.899584] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1181.899584] env[62368]: value = "task-1199307" [ 1181.899584] env[62368]: _type = "Task" [ 1181.899584] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.908755] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199307, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.276506] env[62368]: DEBUG oslo_vmware.api [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Task: {'id': task-1199306, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204312} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.276813] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1182.277018] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1182.277211] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1182.277392] env[62368]: INFO nova.compute.manager [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1182.277642] env[62368]: DEBUG oslo.service.loopingcall [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1182.277840] env[62368]: DEBUG nova.compute.manager [-] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1182.277944] env[62368]: DEBUG nova.network.neutron [-] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1182.409497] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199307, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063267} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.409974] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1182.410579] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33db8fc8-ed73-4805-a5c8-86d155b7f580 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.433124] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] b84d58f8-de40-493f-9586-705ebdc4b759/b84d58f8-de40-493f-9586-705ebdc4b759.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1182.433532] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbff89b6-38e7-4e14-ac9d-b5f85d4bdb24 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.461295] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1182.461295] env[62368]: value = "task-1199308" [ 1182.461295] env[62368]: _type = "Task" [ 1182.461295] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.468217] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199308, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.500748] env[62368]: DEBUG nova.compute.manager [req-b2799e50-9451-48da-9eb1-220b30c624ff req-cc8cd6fb-aeda-4ba7-b465-9a2a5edb8dab service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Received event network-vif-deleted-2b52c97e-6de1-4b5a-9dc4-23348784eacb {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1182.500877] env[62368]: INFO nova.compute.manager [req-b2799e50-9451-48da-9eb1-220b30c624ff req-cc8cd6fb-aeda-4ba7-b465-9a2a5edb8dab service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Neutron deleted interface 2b52c97e-6de1-4b5a-9dc4-23348784eacb; detaching it from the instance and deleting it from the info cache [ 1182.501067] env[62368]: DEBUG nova.network.neutron [req-b2799e50-9451-48da-9eb1-220b30c624ff req-cc8cd6fb-aeda-4ba7-b465-9a2a5edb8dab service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.970851] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199308, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.980382] env[62368]: DEBUG nova.network.neutron [-] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.003718] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-716563b1-9fff-4a51-a648-8b53ab9bcef8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.013314] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda513dc-c1d0-4d7b-85f8-42b935633189 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.036032] env[62368]: DEBUG nova.compute.manager [req-b2799e50-9451-48da-9eb1-220b30c624ff req-cc8cd6fb-aeda-4ba7-b465-9a2a5edb8dab service nova] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Detach interface failed, port_id=2b52c97e-6de1-4b5a-9dc4-23348784eacb, reason: Instance 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1183.472326] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199308, 'name': ReconfigVM_Task, 'duration_secs': 0.76509} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.472792] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Reconfigured VM instance instance-00000074 to attach disk [datastore1] b84d58f8-de40-493f-9586-705ebdc4b759/b84d58f8-de40-493f-9586-705ebdc4b759.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1183.473436] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6fb19230-3501-4227-84bd-803ce99c80b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.480412] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1183.480412] env[62368]: value = "task-1199309" [ 1183.480412] env[62368]: _type = "Task" [ 1183.480412] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.483542] env[62368]: INFO nova.compute.manager [-] [instance: 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70] Took 1.21 seconds to deallocate network for instance. [ 1183.494020] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199309, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.990505] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.990772] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.991041] env[62368]: DEBUG nova.objects.instance [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lazy-loading 'resources' on Instance uuid 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.992064] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199309, 'name': Rename_Task, 'duration_secs': 0.142819} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.992319] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1183.992742] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef9222e3-fbd3-4b1c-934f-6ace44b46975 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.999050] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1183.999050] env[62368]: value = "task-1199310" [ 1183.999050] env[62368]: _type = "Task" [ 1183.999050] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.006846] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199310, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.508999] env[62368]: DEBUG oslo_vmware.api [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199310, 'name': PowerOnVM_Task, 'duration_secs': 0.446692} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.509343] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1184.509799] env[62368]: INFO nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Took 7.50 seconds to spawn the instance on the hypervisor. [ 1184.509799] env[62368]: DEBUG nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.510612] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca230a4b-e8a8-4d31-ab85-ab9ac389bfb2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.549212] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a6f028-d857-4f58-a6c0-1013becd9e64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.557889] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b582bb1e-90ed-4170-bde7-dc4c80fbfa25 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.589916] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9d7308-f211-40a3-80da-13995c3a39e7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.597469] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b08b047-a887-4db2-b7c1-7ca415ed74d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.610854] env[62368]: DEBUG nova.compute.provider_tree [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1185.032022] env[62368]: INFO nova.compute.manager [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Took 12.32 seconds to build instance. [ 1185.113597] env[62368]: DEBUG nova.scheduler.client.report [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1185.534839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-ab967ae9-62b4-4703-ba1a-e276e76002c1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.834s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.619133] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.628s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.637564] env[62368]: INFO nova.scheduler.client.report [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Deleted allocations for instance 0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70 [ 1185.952123] env[62368]: DEBUG nova.compute.manager [req-ac4797b3-58a0-4bd2-900c-fa6093af8bf7 req-3e658b93-ac66-44c8-b1d4-55641a207d48 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Received event network-changed-7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1185.952331] env[62368]: DEBUG nova.compute.manager [req-ac4797b3-58a0-4bd2-900c-fa6093af8bf7 req-3e658b93-ac66-44c8-b1d4-55641a207d48 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Refreshing instance network info cache due to event network-changed-7fc54e0c-c523-43cc-b1a5-4a35cdf98767. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1185.952554] env[62368]: DEBUG oslo_concurrency.lockutils [req-ac4797b3-58a0-4bd2-900c-fa6093af8bf7 req-3e658b93-ac66-44c8-b1d4-55641a207d48 service nova] Acquiring lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.952701] env[62368]: DEBUG oslo_concurrency.lockutils [req-ac4797b3-58a0-4bd2-900c-fa6093af8bf7 req-3e658b93-ac66-44c8-b1d4-55641a207d48 service nova] Acquired lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.952864] env[62368]: DEBUG nova.network.neutron [req-ac4797b3-58a0-4bd2-900c-fa6093af8bf7 req-3e658b93-ac66-44c8-b1d4-55641a207d48 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Refreshing network info cache for port 7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1186.145322] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e5ad02a8-bd2a-4173-8131-c7f14a105ad1 tempest-ServerRescueTestJSON-1955413993 tempest-ServerRescueTestJSON-1955413993-project-member] Lock "0b9e0046-6e0f-42b1-bc4d-a1f5c13bed70" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.016931] env[62368]: DEBUG nova.network.neutron [req-ac4797b3-58a0-4bd2-900c-fa6093af8bf7 req-3e658b93-ac66-44c8-b1d4-55641a207d48 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Updated VIF entry in instance network info cache for port 7fc54e0c-c523-43cc-b1a5-4a35cdf98767. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1187.017328] env[62368]: DEBUG nova.network.neutron [req-ac4797b3-58a0-4bd2-900c-fa6093af8bf7 req-3e658b93-ac66-44c8-b1d4-55641a207d48 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Updating instance_info_cache with network_info: [{"id": "7fc54e0c-c523-43cc-b1a5-4a35cdf98767", "address": "fa:16:3e:81:d9:f6", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fc54e0c-c5", "ovs_interfaceid": "7fc54e0c-c523-43cc-b1a5-4a35cdf98767", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.519971] env[62368]: DEBUG oslo_concurrency.lockutils [req-ac4797b3-58a0-4bd2-900c-fa6093af8bf7 req-3e658b93-ac66-44c8-b1d4-55641a207d48 service nova] Releasing lock "refresh_cache-b84d58f8-de40-493f-9586-705ebdc4b759" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.597730] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.597951] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.861322] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.860032] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.855577] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.859263] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.859422] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1192.367237] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Didn't find any instances for network info cache update. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1192.859916] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.860642] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.860996] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1194.860699] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.363900] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.364179] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.364311] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.364495] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1195.365446] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e326041c-bb0a-4c2a-befa-912823eed9d1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.374496] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d929d6c6-d39e-40f1-8b00-b33f5b01af51 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.390590] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07ec3eb-d528-4cab-900d-f19fd73c7bcf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.396954] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c71fde7-4573-451b-994e-806cdd3fc549 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.425775] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180694MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1195.425878] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.426082] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.451236] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.451501] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance b84d58f8-de40-493f-9586-705ebdc4b759 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.451596] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1196.451717] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1196.487048] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cba0d05-b72a-43b8-8f08-9ed4ceb69d5b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.494521] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3187ec-7291-40e4-be99-79772320009b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.524608] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af49b18-5c17-45b7-8a54-f5b4f53cd6e8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.531270] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372e94f1-b9a7-47af-9e94-6f3ce13e61ee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.543728] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1197.046831] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1197.552073] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1197.552380] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.126s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.547371] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.394839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.395258] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.395360] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.395561] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.395736] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.398057] env[62368]: INFO nova.compute.manager [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Terminating instance [ 1218.399888] env[62368]: DEBUG nova.compute.manager [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1218.400097] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1218.400936] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ff61d1-d0c6-4b57-985c-e22399ba86b1 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.408622] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1218.408848] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0472c5a-fd0b-4994-8da8-efaf583d23d4 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.416008] env[62368]: DEBUG oslo_vmware.api [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1218.416008] env[62368]: value = "task-1199311" [ 1218.416008] env[62368]: _type = "Task" [ 1218.416008] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.423307] env[62368]: DEBUG oslo_vmware.api [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199311, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.926668] env[62368]: DEBUG oslo_vmware.api [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199311, 'name': PowerOffVM_Task, 'duration_secs': 0.282552} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.926973] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1218.927115] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1218.927372] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3538d66e-9b1f-4364-9d58-5d9993f44f14 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.989275] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1218.989522] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1218.989683] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleting the datastore file [datastore2] 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1218.989950] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7288e977-958b-4574-bb0e-3b9057ab1013 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.995696] env[62368]: DEBUG oslo_vmware.api [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for the task: (returnval){ [ 1218.995696] env[62368]: value = "task-1199313" [ 1218.995696] env[62368]: _type = "Task" [ 1218.995696] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.003394] env[62368]: DEBUG oslo_vmware.api [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.505282] env[62368]: DEBUG oslo_vmware.api [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Task: {'id': task-1199313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12997} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.505646] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1219.505711] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1219.505895] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1219.506089] env[62368]: INFO nova.compute.manager [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1219.506369] env[62368]: DEBUG oslo.service.loopingcall [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1219.506584] env[62368]: DEBUG nova.compute.manager [-] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1219.506683] env[62368]: DEBUG nova.network.neutron [-] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1219.917902] env[62368]: DEBUG nova.compute.manager [req-914d7b6d-4623-42ad-b7e8-49367aff3b2c req-60f4a271-e72a-4e60-ad45-35a3a0cb60e1 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Received event network-vif-deleted-7e8eb12d-c67a-4b59-9e66-b3b530e66e75 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1219.918012] env[62368]: INFO nova.compute.manager [req-914d7b6d-4623-42ad-b7e8-49367aff3b2c req-60f4a271-e72a-4e60-ad45-35a3a0cb60e1 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Neutron deleted interface 7e8eb12d-c67a-4b59-9e66-b3b530e66e75; detaching it from the instance and deleting it from the info cache [ 1219.918204] env[62368]: DEBUG nova.network.neutron [req-914d7b6d-4623-42ad-b7e8-49367aff3b2c req-60f4a271-e72a-4e60-ad45-35a3a0cb60e1 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.394276] env[62368]: DEBUG nova.network.neutron [-] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.420612] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8994a621-b0d7-4ba5-b412-acbe1ddfbe88 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.430613] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51391042-b4be-4ac3-bba2-b89fd8e4a674 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.453796] env[62368]: DEBUG nova.compute.manager [req-914d7b6d-4623-42ad-b7e8-49367aff3b2c req-60f4a271-e72a-4e60-ad45-35a3a0cb60e1 service nova] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Detach interface failed, port_id=7e8eb12d-c67a-4b59-9e66-b3b530e66e75, reason: Instance 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1220.897549] env[62368]: INFO nova.compute.manager [-] [instance: 27ff9bed-7c5f-4553-9d8e-8ec78708caf1] Took 1.39 seconds to deallocate network for instance. [ 1221.404046] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1221.404046] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1221.404256] env[62368]: DEBUG nova.objects.instance [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lazy-loading 'resources' on Instance uuid 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1221.950668] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb119cc-1d73-4a64-a166-65306cf41ca0 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.958240] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a045206-715a-434e-88bf-7fde61918638 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.988461] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2da972-a27b-4057-936b-317ae5f68a85 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.996018] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2bc516f-8ec0-43be-ad81-75663fafef2c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.008575] env[62368]: DEBUG nova.compute.provider_tree [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1222.511786] env[62368]: DEBUG nova.scheduler.client.report [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1223.017581] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.613s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.037600] env[62368]: INFO nova.scheduler.client.report [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Deleted allocations for instance 27ff9bed-7c5f-4553-9d8e-8ec78708caf1 [ 1223.546050] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d23c6d0a-7966-4431-bcb1-13682884fbab tempest-AttachVolumeShelveTestJSON-1823897722 tempest-AttachVolumeShelveTestJSON-1823897722-project-member] Lock "27ff9bed-7c5f-4553-9d8e-8ec78708caf1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.150s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.460765] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b84d58f8-de40-493f-9586-705ebdc4b759" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.461064] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.965019] env[62368]: DEBUG nova.compute.utils [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1225.468180] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.526328] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b84d58f8-de40-493f-9586-705ebdc4b759" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.526704] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.526868] env[62368]: INFO nova.compute.manager [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Attaching volume 780e03d7-8828-4f72-bf19-8bc59bdd6c3d to /dev/sdb [ 1226.555959] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b1c42a-31c0-450a-afa8-88195e2e06ec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.563440] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a64d9e-0d12-4059-ac0f-ce08985cb58a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.576038] env[62368]: DEBUG nova.virt.block_device [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Updating existing volume attachment record: bb113965-7c2f-4534-9f6d-dc282045ff65 {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1231.117623] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1231.117623] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259930', 'volume_id': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'name': 'volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b84d58f8-de40-493f-9586-705ebdc4b759', 'attached_at': '', 'detached_at': '', 'volume_id': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'serial': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1231.118277] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c97b56a-4102-4cb2-ad59-a24f8be98eac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.135945] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29452c6-102e-45cc-80df-e3f7e677465c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.159411] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d/volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1231.159667] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb59ee31-cd4b-44d3-a67b-75069f180044 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.176961] env[62368]: DEBUG oslo_vmware.api [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1231.176961] env[62368]: value = "task-1199317" [ 1231.176961] env[62368]: _type = "Task" [ 1231.176961] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.184313] env[62368]: DEBUG oslo_vmware.api [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199317, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.687057] env[62368]: DEBUG oslo_vmware.api [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199317, 'name': ReconfigVM_Task, 'duration_secs': 0.32115} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.687347] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Reconfigured VM instance instance-00000074 to attach disk [datastore2] volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d/volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1231.692160] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbf95df7-a5fb-49c1-87e4-76f03095b3f5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.706853] env[62368]: DEBUG oslo_vmware.api [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1231.706853] env[62368]: value = "task-1199318" [ 1231.706853] env[62368]: _type = "Task" [ 1231.706853] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.713944] env[62368]: DEBUG oslo_vmware.api [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199318, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.217047] env[62368]: DEBUG oslo_vmware.api [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199318, 'name': ReconfigVM_Task, 'duration_secs': 0.142432} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.217336] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259930', 'volume_id': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'name': 'volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b84d58f8-de40-493f-9586-705ebdc4b759', 'attached_at': '', 'detached_at': '', 'volume_id': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'serial': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1233.252462] env[62368]: DEBUG nova.objects.instance [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'flavor' on Instance uuid b84d58f8-de40-493f-9586-705ebdc4b759 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1233.757558] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cb29c9c1-7f33-45b5-bce2-aea98f3b4066 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.231s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.980514] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b84d58f8-de40-493f-9586-705ebdc4b759" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.980772] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.483968] env[62368]: INFO nova.compute.manager [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Detaching volume 780e03d7-8828-4f72-bf19-8bc59bdd6c3d [ 1234.513533] env[62368]: INFO nova.virt.block_device [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Attempting to driver detach volume 780e03d7-8828-4f72-bf19-8bc59bdd6c3d from mountpoint /dev/sdb [ 1234.513783] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1234.513972] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259930', 'volume_id': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'name': 'volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b84d58f8-de40-493f-9586-705ebdc4b759', 'attached_at': '', 'detached_at': '', 'volume_id': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'serial': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1234.514892] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc89626-f1d8-442a-8278-294dac95f110 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.536050] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9af626-6843-49ed-a840-5ed558a9ca2e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.542331] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0413ff-302a-4912-a815-80beda914e49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.561432] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a94abf-effd-42b3-a5d3-06a1960a9342 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.575861] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] The volume has not been displaced from its original location: [datastore2] volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d/volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1234.581055] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Reconfiguring VM instance instance-00000074 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1234.581305] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7130f92-5cea-415f-812f-3b783b03f40b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.598637] env[62368]: DEBUG oslo_vmware.api [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1234.598637] env[62368]: value = "task-1199319" [ 1234.598637] env[62368]: _type = "Task" [ 1234.598637] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.605745] env[62368]: DEBUG oslo_vmware.api [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.109903] env[62368]: DEBUG oslo_vmware.api [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199319, 'name': ReconfigVM_Task, 'duration_secs': 0.206824} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.110139] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Reconfigured VM instance instance-00000074 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1235.114618] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8fc99a93-0890-42f7-8857-37e170ba4e05 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.130639] env[62368]: DEBUG oslo_vmware.api [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1235.130639] env[62368]: value = "task-1199320" [ 1235.130639] env[62368]: _type = "Task" [ 1235.130639] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.138052] env[62368]: DEBUG oslo_vmware.api [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199320, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.640228] env[62368]: DEBUG oslo_vmware.api [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199320, 'name': ReconfigVM_Task, 'duration_secs': 0.173837} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.640567] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259930', 'volume_id': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'name': 'volume-780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b84d58f8-de40-493f-9586-705ebdc4b759', 'attached_at': '', 'detached_at': '', 'volume_id': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d', 'serial': '780e03d7-8828-4f72-bf19-8bc59bdd6c3d'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1236.179874] env[62368]: DEBUG nova.objects.instance [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'flavor' on Instance uuid b84d58f8-de40-493f-9586-705ebdc4b759 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1237.188199] env[62368]: DEBUG oslo_concurrency.lockutils [None req-b471666f-a755-4c13-a8dc-50b21f19224f tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.207s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.211203] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b84d58f8-de40-493f-9586-705ebdc4b759" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.211203] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.211586] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "b84d58f8-de40-493f-9586-705ebdc4b759-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.211586] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.211661] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.213847] env[62368]: INFO nova.compute.manager [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Terminating instance [ 1238.215641] env[62368]: DEBUG nova.compute.manager [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1238.215873] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1238.216736] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c323221c-9a12-4ba5-9d42-2214a5dc5c36 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.224603] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1238.224851] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01652c78-190d-4918-9415-1550f91512ae {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.230940] env[62368]: DEBUG oslo_vmware.api [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1238.230940] env[62368]: value = "task-1199321" [ 1238.230940] env[62368]: _type = "Task" [ 1238.230940] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.238440] env[62368]: DEBUG oslo_vmware.api [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199321, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.741396] env[62368]: DEBUG oslo_vmware.api [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199321, 'name': PowerOffVM_Task, 'duration_secs': 0.180794} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.741785] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1238.742055] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1238.742368] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d319ba31-7bec-48db-a395-d4da4b689e64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.807258] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1238.807490] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Deleting contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1238.807676] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleting the datastore file [datastore1] b84d58f8-de40-493f-9586-705ebdc4b759 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1238.807941] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6916b3a-2f42-4e74-971f-0efae7336181 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.814275] env[62368]: DEBUG oslo_vmware.api [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1238.814275] env[62368]: value = "task-1199323" [ 1238.814275] env[62368]: _type = "Task" [ 1238.814275] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.821526] env[62368]: DEBUG oslo_vmware.api [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199323, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.325530] env[62368]: DEBUG oslo_vmware.api [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199323, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135275} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.325893] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1239.326065] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Deleted contents of the VM from datastore datastore1 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1239.326268] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1239.326456] env[62368]: INFO nova.compute.manager [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1239.326705] env[62368]: DEBUG oslo.service.loopingcall [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1239.326909] env[62368]: DEBUG nova.compute.manager [-] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1239.327014] env[62368]: DEBUG nova.network.neutron [-] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1239.726012] env[62368]: DEBUG nova.compute.manager [req-316c0661-d77c-422c-bb99-b4f51c480d32 req-64bf9120-4dac-4beb-8788-7f77ac3a9080 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Received event network-vif-deleted-7fc54e0c-c523-43cc-b1a5-4a35cdf98767 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1239.726338] env[62368]: INFO nova.compute.manager [req-316c0661-d77c-422c-bb99-b4f51c480d32 req-64bf9120-4dac-4beb-8788-7f77ac3a9080 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Neutron deleted interface 7fc54e0c-c523-43cc-b1a5-4a35cdf98767; detaching it from the instance and deleting it from the info cache [ 1239.726621] env[62368]: DEBUG nova.network.neutron [req-316c0661-d77c-422c-bb99-b4f51c480d32 req-64bf9120-4dac-4beb-8788-7f77ac3a9080 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.204575] env[62368]: DEBUG nova.network.neutron [-] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.229092] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3ce7c34-c1b5-49cb-9c5c-1d72196c2545 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.238751] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89e724e-3b65-410b-b3d7-7f509248c9bd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.262705] env[62368]: DEBUG nova.compute.manager [req-316c0661-d77c-422c-bb99-b4f51c480d32 req-64bf9120-4dac-4beb-8788-7f77ac3a9080 service nova] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Detach interface failed, port_id=7fc54e0c-c523-43cc-b1a5-4a35cdf98767, reason: Instance b84d58f8-de40-493f-9586-705ebdc4b759 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1240.708591] env[62368]: INFO nova.compute.manager [-] [instance: b84d58f8-de40-493f-9586-705ebdc4b759] Took 1.38 seconds to deallocate network for instance. [ 1241.215118] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.215401] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1241.215639] env[62368]: DEBUG nova.objects.instance [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'resources' on Instance uuid b84d58f8-de40-493f-9586-705ebdc4b759 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1241.751051] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e321c19-6b35-4309-b70b-391275b008d2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.757896] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f88d89-95e5-4eb8-853c-a154b53e6be6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.787086] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c45b741-e0a1-4233-92c8-a001d0ca51b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.793585] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0b5b9c-85f1-44b6-9c06-c9388937bee9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.807230] env[62368]: DEBUG nova.compute.provider_tree [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1242.332456] env[62368]: ERROR nova.scheduler.client.report [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [req-c3bc2e4c-f9cd-41e0-a365-3b343405c971] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c3bc2e4c-f9cd-41e0-a365-3b343405c971"}]} [ 1242.353844] env[62368]: DEBUG nova.scheduler.client.report [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1242.369903] env[62368]: DEBUG nova.scheduler.client.report [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1242.370158] env[62368]: DEBUG nova.compute.provider_tree [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1242.380467] env[62368]: DEBUG nova.scheduler.client.report [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1242.397599] env[62368]: DEBUG nova.scheduler.client.report [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1242.420066] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a482d19-3d67-42ce-b14d-90c69a1449fa {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.427445] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5302fa-ca09-400c-89aa-fe1ef3572dec {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.456242] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d74137-2ec8-4111-84c3-77b2c78038e2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.462737] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed97f1bb-9cf0-4c1f-8dda-31126b69bfdd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.475034] env[62368]: DEBUG nova.compute.provider_tree [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1243.491902] env[62368]: DEBUG nova.scheduler.client.report [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 157 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1243.492236] env[62368]: DEBUG nova.compute.provider_tree [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 157 to 158 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1243.492396] env[62368]: DEBUG nova.compute.provider_tree [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1243.996828] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.781s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.018850] env[62368]: INFO nova.scheduler.client.report [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleted allocations for instance b84d58f8-de40-493f-9586-705ebdc4b759 [ 1244.526328] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e46c6796-0909-4e3e-a71c-66e737e485e4 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "b84d58f8-de40-493f-9586-705ebdc4b759" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.315s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.859722] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.223035] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.223035] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.724968] env[62368]: DEBUG nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Starting instance... {{(pid=62368) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1246.860748] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.243569] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.243820] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.245348] env[62368]: INFO nova.compute.claims [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1248.278261] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0301eba-7b1d-4057-934b-13ad77bbb7b9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.285969] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe68d3d-9459-4198-b65d-16244b046e82 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.314685] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1021af6e-4c12-43e9-a919-7fa1b7dc7eac {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.321236] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5cf72e-7a86-4a4f-8941-3817c2cdf872 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.334854] env[62368]: DEBUG nova.compute.provider_tree [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1248.837493] env[62368]: DEBUG nova.scheduler.client.report [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1249.342320] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.098s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.342853] env[62368]: DEBUG nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Start building networks asynchronously for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1249.847917] env[62368]: DEBUG nova.compute.utils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1249.849434] env[62368]: DEBUG nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Allocating IP information in the background. {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1249.849610] env[62368]: DEBUG nova.network.neutron [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] allocate_for_instance() {{(pid=62368) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1249.859940] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.897849] env[62368]: DEBUG nova.policy [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac2df24b03d147f5a33d67a97bbb5bb2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e41fb7a4acf043df959806368ae94ef9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62368) authorize /opt/stack/nova/nova/policy.py:201}} [ 1250.179597] env[62368]: DEBUG nova.network.neutron [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Successfully created port: 8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1250.353247] env[62368]: DEBUG nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Start building block device mappings for instance. {{(pid=62368) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1250.856600] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.861109] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.861265] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1250.861387] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Rebuilding the list of instances to heal {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1251.364161] env[62368]: DEBUG nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Start spawning the instance on the hypervisor. {{(pid=62368) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1251.366751] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Skipping network cache update for instance because it is Building. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1251.366920] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Didn't find any instances for network info cache update. {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1251.389952] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-08T23:11:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-08T23:10:52Z,direct_url=,disk_format='vmdk',id=d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2a054e975862463bb842479ad19393ca',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-08T23:10:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1251.390214] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Flavor limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1251.390377] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Image limits 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1251.390566] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Flavor pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1251.390716] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Image pref 0:0:0 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1251.390868] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62368) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1251.391095] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1251.391261] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1251.391432] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Got 1 possible topologies {{(pid=62368) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1251.391595] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1251.391767] env[62368]: DEBUG nova.virt.hardware [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62368) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1251.392645] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71e13fa-0317-4f3f-b575-160273c4f3eb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.400836] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a85c860-8eb3-4064-b5b4-0f1cb8e9ddfb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.541143] env[62368]: DEBUG nova.compute.manager [req-f91b284e-f91c-413d-9a73-a9a1ee5982b9 req-d9feec10-04a9-4b4b-a624-6f258be71fcb service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Received event network-vif-plugged-8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1251.541287] env[62368]: DEBUG oslo_concurrency.lockutils [req-f91b284e-f91c-413d-9a73-a9a1ee5982b9 req-d9feec10-04a9-4b4b-a624-6f258be71fcb service nova] Acquiring lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.541491] env[62368]: DEBUG oslo_concurrency.lockutils [req-f91b284e-f91c-413d-9a73-a9a1ee5982b9 req-d9feec10-04a9-4b4b-a624-6f258be71fcb service nova] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.541656] env[62368]: DEBUG oslo_concurrency.lockutils [req-f91b284e-f91c-413d-9a73-a9a1ee5982b9 req-d9feec10-04a9-4b4b-a624-6f258be71fcb service nova] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.541831] env[62368]: DEBUG nova.compute.manager [req-f91b284e-f91c-413d-9a73-a9a1ee5982b9 req-d9feec10-04a9-4b4b-a624-6f258be71fcb service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] No waiting events found dispatching network-vif-plugged-8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1251.542043] env[62368]: WARNING nova.compute.manager [req-f91b284e-f91c-413d-9a73-a9a1ee5982b9 req-d9feec10-04a9-4b4b-a624-6f258be71fcb service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Received unexpected event network-vif-plugged-8cc63649-ae23-4a32-8e8c-3b311d8a5d91 for instance with vm_state building and task_state spawning. [ 1251.624619] env[62368]: DEBUG nova.network.neutron [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Successfully updated port: 8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1251.859937] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1252.130323] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1252.130467] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.130621] env[62368]: DEBUG nova.network.neutron [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Building network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1252.663415] env[62368]: DEBUG nova.network.neutron [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1252.779187] env[62368]: DEBUG nova.network.neutron [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updating instance_info_cache with network_info: [{"id": "8cc63649-ae23-4a32-8e8c-3b311d8a5d91", "address": "fa:16:3e:b4:cf:b9", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cc63649-ae", "ovs_interfaceid": "8cc63649-ae23-4a32-8e8c-3b311d8a5d91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.859946] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1253.281444] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1253.281772] env[62368]: DEBUG nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Instance network_info: |[{"id": "8cc63649-ae23-4a32-8e8c-3b311d8a5d91", "address": "fa:16:3e:b4:cf:b9", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cc63649-ae", "ovs_interfaceid": "8cc63649-ae23-4a32-8e8c-3b311d8a5d91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62368) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1253.282226] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:cf:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8cc63649-ae23-4a32-8e8c-3b311d8a5d91', 'vif_model': 'vmxnet3'}] {{(pid=62368) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1253.289600] env[62368]: DEBUG oslo.service.loopingcall [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1253.289829] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Creating VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1253.290070] env[62368]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-630b9528-2446-4318-a2ba-6ef9311c7acd {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.309578] env[62368]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1253.309578] env[62368]: value = "task-1199324" [ 1253.309578] env[62368]: _type = "Task" [ 1253.309578] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.316647] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199324, 'name': CreateVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.567497] env[62368]: DEBUG nova.compute.manager [req-0d28b8c0-e6bf-46b6-9794-68d098ac0b45 req-aee52d07-20c6-4f50-a247-3fb810df6e90 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Received event network-changed-8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1253.567709] env[62368]: DEBUG nova.compute.manager [req-0d28b8c0-e6bf-46b6-9794-68d098ac0b45 req-aee52d07-20c6-4f50-a247-3fb810df6e90 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Refreshing instance network info cache due to event network-changed-8cc63649-ae23-4a32-8e8c-3b311d8a5d91. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1253.567955] env[62368]: DEBUG oslo_concurrency.lockutils [req-0d28b8c0-e6bf-46b6-9794-68d098ac0b45 req-aee52d07-20c6-4f50-a247-3fb810df6e90 service nova] Acquiring lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1253.568294] env[62368]: DEBUG oslo_concurrency.lockutils [req-0d28b8c0-e6bf-46b6-9794-68d098ac0b45 req-aee52d07-20c6-4f50-a247-3fb810df6e90 service nova] Acquired lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.568502] env[62368]: DEBUG nova.network.neutron [req-0d28b8c0-e6bf-46b6-9794-68d098ac0b45 req-aee52d07-20c6-4f50-a247-3fb810df6e90 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Refreshing network info cache for port 8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1253.819568] env[62368]: DEBUG oslo_vmware.api [-] Task: {'id': task-1199324, 'name': CreateVM_Task, 'duration_secs': 0.289518} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.819919] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Created VM on the ESX host {{(pid=62368) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1253.820396] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1253.820564] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.820913] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1253.821164] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4b959df-aa13-4974-a8cb-77651a376d83 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.825318] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1253.825318] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52312a10-b83b-7661-a5fd-18d91519ff04" [ 1253.825318] env[62368]: _type = "Task" [ 1253.825318] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.832548] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52312a10-b83b-7661-a5fd-18d91519ff04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.249979] env[62368]: DEBUG nova.network.neutron [req-0d28b8c0-e6bf-46b6-9794-68d098ac0b45 req-aee52d07-20c6-4f50-a247-3fb810df6e90 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updated VIF entry in instance network info cache for port 8cc63649-ae23-4a32-8e8c-3b311d8a5d91. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1254.250578] env[62368]: DEBUG nova.network.neutron [req-0d28b8c0-e6bf-46b6-9794-68d098ac0b45 req-aee52d07-20c6-4f50-a247-3fb810df6e90 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updating instance_info_cache with network_info: [{"id": "8cc63649-ae23-4a32-8e8c-3b311d8a5d91", "address": "fa:16:3e:b4:cf:b9", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cc63649-ae", "ovs_interfaceid": "8cc63649-ae23-4a32-8e8c-3b311d8a5d91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1254.335770] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52312a10-b83b-7661-a5fd-18d91519ff04, 'name': SearchDatastore_Task, 'duration_secs': 0.01014} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.336065] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1254.336306] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Processing image d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d {{(pid=62368) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1254.336538] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1254.336689] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.336871] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1254.337146] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a183338-3a5e-4cae-b985-ad1ab68c79ea {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.344629] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62368) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1254.344802] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62368) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1254.345553] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08d42061-153d-4809-8965-9eba720996d6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.350157] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1254.350157] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52dd937f-680d-a22e-6f7d-90ee75f14076" [ 1254.350157] env[62368]: _type = "Task" [ 1254.350157] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.356999] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52dd937f-680d-a22e-6f7d-90ee75f14076, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.753715] env[62368]: DEBUG oslo_concurrency.lockutils [req-0d28b8c0-e6bf-46b6-9794-68d098ac0b45 req-aee52d07-20c6-4f50-a247-3fb810df6e90 service nova] Releasing lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1254.860063] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52dd937f-680d-a22e-6f7d-90ee75f14076, 'name': SearchDatastore_Task, 'duration_secs': 0.007306} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.860811] env[62368]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd5e2852-e6ef-4b40-89c9-759de92de484 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.865422] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1254.865422] env[62368]: value = "session[529ee714-52b3-30cd-d808-84351b384cd1]52fe14cd-ffad-d441-ac6b-6f6438821f33" [ 1254.865422] env[62368]: _type = "Task" [ 1254.865422] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.872435] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fe14cd-ffad-d441-ac6b-6f6438821f33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.377092] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': session[529ee714-52b3-30cd-d808-84351b384cd1]52fe14cd-ffad-d441-ac6b-6f6438821f33, 'name': SearchDatastore_Task, 'duration_secs': 0.0089} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.377353] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1255.377603] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 8894539b-84c5-4fee-95e3-6a4ff1f15d53/8894539b-84c5-4fee-95e3-6a4ff1f15d53.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1255.377856] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c479d672-7e9a-4d31-91d7-6bd2009b2519 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.385218] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1255.385218] env[62368]: value = "task-1199325" [ 1255.385218] env[62368]: _type = "Task" [ 1255.385218] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.392490] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199325, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.860258] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.860708] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1255.860708] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.895012] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199325, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42607} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.895275] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d/d6bd1b2e-f5f9-49c5-bb6f-9536d441b77d.vmdk to [datastore2] 8894539b-84c5-4fee-95e3-6a4ff1f15d53/8894539b-84c5-4fee-95e3-6a4ff1f15d53.vmdk {{(pid=62368) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1255.895482] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Extending root virtual disk to 1048576 {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1255.895716] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5c1daca3-e47c-428d-ab67-52edde77619c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.902145] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1255.902145] env[62368]: value = "task-1199326" [ 1255.902145] env[62368]: _type = "Task" [ 1255.902145] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.909234] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199326, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.363309] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1256.363666] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.363745] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.363853] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1256.364770] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c0aa73-a852-44a9-9004-6c82fd23423e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.373403] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f311f79-529c-4da4-9788-6610e80348b2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.387011] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3dda459-3950-47cc-8978-44b8df069f2d {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.393687] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6220640-bb26-4115-9c7a-60cf78089dd2 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.421281] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181026MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1256.421411] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1256.421594] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.430414] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199326, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055681} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.430646] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Extended root virtual disk {{(pid=62368) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1256.431350] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1ec7b2-3f76-4dd3-853f-73d4bfc7b353 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.453119] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] 8894539b-84c5-4fee-95e3-6a4ff1f15d53/8894539b-84c5-4fee-95e3-6a4ff1f15d53.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1256.453584] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43499285-a51b-4a91-a286-b040799d3a60 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.472465] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1256.472465] env[62368]: value = "task-1199327" [ 1256.472465] env[62368]: _type = "Task" [ 1256.472465] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.480339] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199327, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.982817] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199327, 'name': ReconfigVM_Task, 'duration_secs': 0.270567} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.983212] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Reconfigured VM instance instance-00000075 to attach disk [datastore2] 8894539b-84c5-4fee-95e3-6a4ff1f15d53/8894539b-84c5-4fee-95e3-6a4ff1f15d53.vmdk or device None with type sparse {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1256.983785] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09dc01c4-b40a-4c5d-9260-a387e756e7f6 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.989836] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1256.989836] env[62368]: value = "task-1199328" [ 1256.989836] env[62368]: _type = "Task" [ 1256.989836] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.997437] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199328, 'name': Rename_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.449816] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Instance 8894539b-84c5-4fee-95e3-6a4ff1f15d53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62368) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1257.450055] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1257.450204] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1257.476339] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b459e75-dae0-42f5-81d6-ceda5cb4e98e {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.483960] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c91c67-add8-4f08-9fe6-61ce5abbc46c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.514777] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d5bef9-c469-44b5-b68a-0574a032bee7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.522398] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199328, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.525196] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8add6583-e48b-4085-a5f2-f2071d9920f3 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.538061] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1258.022190] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199328, 'name': Rename_Task} progress is 99%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.057611] env[62368]: ERROR nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [req-be7c3305-7623-41ee-a751-ba6ada9ec531] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2202a74c-753d-4e1d-a031-7cefe24ee9d6. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-be7c3305-7623-41ee-a751-ba6ada9ec531"}]} [ 1258.073732] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Refreshing inventories for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1258.088028] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating ProviderTree inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1258.088225] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 157, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1258.100200] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Refreshing aggregate associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, aggregates: None {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1258.120274] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Refreshing trait associations for resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62368) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1258.147344] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d62de1-f607-4e4e-b5d1-ed7eda1936c7 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.156756] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4de8111-93c7-4e3e-89d8-921799c90515 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.186684] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416d7ab7-fe09-4c03-9857-0148b36c990a {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.194457] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7a1b7a-84da-48cc-802b-712ce5e2d401 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.207969] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1258.523036] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199328, 'name': Rename_Task, 'duration_secs': 1.146231} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.523296] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Powering on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1258.523515] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6fc4860e-84e8-429f-9050-0d142c9d480b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.530478] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1258.530478] env[62368]: value = "task-1199329" [ 1258.530478] env[62368]: _type = "Task" [ 1258.530478] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.538265] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199329, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.737185] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updated inventory for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with generation 159 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1258.737436] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating resource provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 generation from 159 to 160 during operation: update_inventory {{(pid=62368) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1258.737587] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Updating inventory in ProviderTree for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1259.040135] env[62368]: DEBUG oslo_vmware.api [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199329, 'name': PowerOnVM_Task, 'duration_secs': 0.503251} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.040564] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Powered on the VM {{(pid=62368) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1259.040633] env[62368]: INFO nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Took 7.68 seconds to spawn the instance on the hypervisor. [ 1259.040888] env[62368]: DEBUG nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Checking state {{(pid=62368) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1259.041590] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5afdc1-15a7-4aca-b009-2a3cd371655f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.242335] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1259.242542] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.821s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.561887] env[62368]: INFO nova.compute.manager [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Took 12.33 seconds to build instance. [ 1260.064788] env[62368]: DEBUG oslo_concurrency.lockutils [None req-76f6cc61-093a-4fab-85fb-4f843179fa6d tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.842s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1260.555201] env[62368]: DEBUG nova.compute.manager [req-e134e31d-42c3-40ad-890d-910cce176558 req-78611326-a12a-4c7d-82af-4bf8eb680fd2 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Received event network-changed-8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1260.555401] env[62368]: DEBUG nova.compute.manager [req-e134e31d-42c3-40ad-890d-910cce176558 req-78611326-a12a-4c7d-82af-4bf8eb680fd2 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Refreshing instance network info cache due to event network-changed-8cc63649-ae23-4a32-8e8c-3b311d8a5d91. {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1260.555615] env[62368]: DEBUG oslo_concurrency.lockutils [req-e134e31d-42c3-40ad-890d-910cce176558 req-78611326-a12a-4c7d-82af-4bf8eb680fd2 service nova] Acquiring lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1260.555762] env[62368]: DEBUG oslo_concurrency.lockutils [req-e134e31d-42c3-40ad-890d-910cce176558 req-78611326-a12a-4c7d-82af-4bf8eb680fd2 service nova] Acquired lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1260.555925] env[62368]: DEBUG nova.network.neutron [req-e134e31d-42c3-40ad-890d-910cce176558 req-78611326-a12a-4c7d-82af-4bf8eb680fd2 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Refreshing network info cache for port 8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1261.256630] env[62368]: DEBUG nova.network.neutron [req-e134e31d-42c3-40ad-890d-910cce176558 req-78611326-a12a-4c7d-82af-4bf8eb680fd2 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updated VIF entry in instance network info cache for port 8cc63649-ae23-4a32-8e8c-3b311d8a5d91. {{(pid=62368) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1261.257067] env[62368]: DEBUG nova.network.neutron [req-e134e31d-42c3-40ad-890d-910cce176558 req-78611326-a12a-4c7d-82af-4bf8eb680fd2 service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updating instance_info_cache with network_info: [{"id": "8cc63649-ae23-4a32-8e8c-3b311d8a5d91", "address": "fa:16:3e:b4:cf:b9", "network": {"id": "43011678-a4be-4d6c-b1d2-15c21c65f60e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1610044722-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e41fb7a4acf043df959806368ae94ef9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cc63649-ae", "ovs_interfaceid": "8cc63649-ae23-4a32-8e8c-3b311d8a5d91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1261.760250] env[62368]: DEBUG oslo_concurrency.lockutils [req-e134e31d-42c3-40ad-890d-910cce176558 req-78611326-a12a-4c7d-82af-4bf8eb680fd2 service nova] Releasing lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1296.605219] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1296.605523] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1297.108761] env[62368]: DEBUG nova.compute.utils [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Using /dev/sd instead of None {{(pid=62368) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1297.612221] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.671752] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1298.672082] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1298.672296] env[62368]: INFO nova.compute.manager [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Attaching volume eb0c3075-304a-4c9e-a60c-6c5e3b2cf024 to /dev/sdb [ 1298.702513] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19fd372-3c6d-4ebe-83d1-9fb89fd14aee {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.709645] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc5036e-70e0-4842-8ead-840df36ea527 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.722368] env[62368]: DEBUG nova.virt.block_device [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updating existing volume attachment record: 84668ede-9764-4ed6-95e2-13e27904516c {{(pid=62368) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1303.264820] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Volume attach. Driver type: vmdk {{(pid=62368) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1303.265114] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259932', 'volume_id': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'name': 'volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8894539b-84c5-4fee-95e3-6a4ff1f15d53', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'serial': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1303.265966] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95ec6d4-d918-42ec-8eb7-23acc802bc5f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.282441] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3708b42f-5f6e-4239-bbea-4972b7bfd376 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.306059] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024/volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1303.306293] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f2a1e12-3a61-441f-96e1-dc74c9773a57 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.323826] env[62368]: DEBUG oslo_vmware.api [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1303.323826] env[62368]: value = "task-1199332" [ 1303.323826] env[62368]: _type = "Task" [ 1303.323826] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.331147] env[62368]: DEBUG oslo_vmware.api [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199332, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.833623] env[62368]: DEBUG oslo_vmware.api [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199332, 'name': ReconfigVM_Task, 'duration_secs': 0.336947} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.833887] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Reconfigured VM instance instance-00000075 to attach disk [datastore1] volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024/volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024.vmdk or device None with type thin {{(pid=62368) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1303.838517] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07961140-c1c8-417c-aafd-2d1d5275bacf {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.852766] env[62368]: DEBUG oslo_vmware.api [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1303.852766] env[62368]: value = "task-1199333" [ 1303.852766] env[62368]: _type = "Task" [ 1303.852766] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.859980] env[62368]: DEBUG oslo_vmware.api [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.362816] env[62368]: DEBUG oslo_vmware.api [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199333, 'name': ReconfigVM_Task, 'duration_secs': 0.135383} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.363114] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259932', 'volume_id': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'name': 'volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8894539b-84c5-4fee-95e3-6a4ff1f15d53', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'serial': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024'} {{(pid=62368) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1305.398168] env[62368]: DEBUG nova.objects.instance [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'flavor' on Instance uuid 8894539b-84c5-4fee-95e3-6a4ff1f15d53 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1305.904310] env[62368]: DEBUG oslo_concurrency.lockutils [None req-d8e70639-94cf-4b84-9c4c-c20d42f15b4c tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.232s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.081025] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1306.081276] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1306.583980] env[62368]: INFO nova.compute.manager [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Detaching volume eb0c3075-304a-4c9e-a60c-6c5e3b2cf024 [ 1306.613762] env[62368]: INFO nova.virt.block_device [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Attempting to driver detach volume eb0c3075-304a-4c9e-a60c-6c5e3b2cf024 from mountpoint /dev/sdb [ 1306.614013] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Volume detach. Driver type: vmdk {{(pid=62368) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1306.614219] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259932', 'volume_id': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'name': 'volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8894539b-84c5-4fee-95e3-6a4ff1f15d53', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'serial': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1306.615093] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71de7849-45c8-4d85-b2db-992def2bc980 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.636486] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b247276-4746-47da-b75f-85f43d0760ce {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.643455] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a68291e-5d6e-4611-b42e-8c3c2a65ef23 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.663252] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51de530d-b98b-4808-9a10-48055721ffb8 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.678674] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] The volume has not been displaced from its original location: [datastore1] volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024/volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024.vmdk. No consolidation needed. {{(pid=62368) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1306.683964] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Reconfiguring VM instance instance-00000075 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1306.684231] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa6a90fa-7e4c-451d-a3fa-dfcb5f6ddc55 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.701963] env[62368]: DEBUG oslo_vmware.api [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1306.701963] env[62368]: value = "task-1199334" [ 1306.701963] env[62368]: _type = "Task" [ 1306.701963] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.709172] env[62368]: DEBUG oslo_vmware.api [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199334, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.211110] env[62368]: DEBUG oslo_vmware.api [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199334, 'name': ReconfigVM_Task, 'duration_secs': 0.214837} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.211397] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Reconfigured VM instance instance-00000075 to detach disk 2001 {{(pid=62368) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1307.215964] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a69d7cf7-ed2e-4dc9-8ea0-a2946a63784b {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.230175] env[62368]: DEBUG oslo_vmware.api [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1307.230175] env[62368]: value = "task-1199335" [ 1307.230175] env[62368]: _type = "Task" [ 1307.230175] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.237485] env[62368]: DEBUG oslo_vmware.api [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199335, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.739683] env[62368]: DEBUG oslo_vmware.api [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199335, 'name': ReconfigVM_Task, 'duration_secs': 0.133478} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.740018] env[62368]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-259932', 'volume_id': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'name': 'volume-eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8894539b-84c5-4fee-95e3-6a4ff1f15d53', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024', 'serial': 'eb0c3075-304a-4c9e-a60c-6c5e3b2cf024'} {{(pid=62368) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1308.280546] env[62368]: DEBUG nova.objects.instance [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'flavor' on Instance uuid 8894539b-84c5-4fee-95e3-6a4ff1f15d53 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1309.242499] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1309.242886] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1309.287542] env[62368]: DEBUG oslo_concurrency.lockutils [None req-e1c376f6-2c90-48d4-948a-3924c3e968e1 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.206s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1310.314272] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1310.314677] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1310.314911] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1310.315199] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1310.315398] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1310.317649] env[62368]: INFO nova.compute.manager [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Terminating instance [ 1310.319480] env[62368]: DEBUG nova.compute.manager [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Start destroying the instance on the hypervisor. {{(pid=62368) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1310.319674] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Destroying instance {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1310.320525] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3f2fc8-60f4-4d5c-bc83-3a08294263c5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.328671] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Powering off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1310.329271] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a39a6b7-23ce-4095-b938-c13305956f2f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.336095] env[62368]: DEBUG oslo_vmware.api [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1310.336095] env[62368]: value = "task-1199336" [ 1310.336095] env[62368]: _type = "Task" [ 1310.336095] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.347817] env[62368]: DEBUG oslo_vmware.api [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199336, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.846213] env[62368]: DEBUG oslo_vmware.api [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199336, 'name': PowerOffVM_Task, 'duration_secs': 0.168702} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.846478] env[62368]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Powered off the VM {{(pid=62368) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1310.846634] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Unregistering the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1310.846876] env[62368]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f51ec2bb-3888-4a37-8b45-e90178e24461 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.860194] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1310.914347] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Unregistered the VM {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1310.914579] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Deleting contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1310.914788] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleting the datastore file [datastore2] 8894539b-84c5-4fee-95e3-6a4ff1f15d53 {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1310.915069] env[62368]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfec6727-9aef-4738-ad72-8e6313309b4c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.921852] env[62368]: DEBUG oslo_vmware.api [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for the task: (returnval){ [ 1310.921852] env[62368]: value = "task-1199338" [ 1310.921852] env[62368]: _type = "Task" [ 1310.921852] env[62368]: } to complete. {{(pid=62368) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.928947] env[62368]: DEBUG oslo_vmware.api [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.431962] env[62368]: DEBUG oslo_vmware.api [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Task: {'id': task-1199338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133841} completed successfully. {{(pid=62368) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.432422] env[62368]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleted the datastore file {{(pid=62368) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1311.432422] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Deleted contents of the VM from datastore datastore2 {{(pid=62368) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1311.432616] env[62368]: DEBUG nova.virt.vmwareapi.vmops [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Instance destroyed {{(pid=62368) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1311.432795] env[62368]: INFO nova.compute.manager [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1311.433048] env[62368]: DEBUG oslo.service.loopingcall [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62368) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1311.433248] env[62368]: DEBUG nova.compute.manager [-] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Deallocating network for instance {{(pid=62368) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1311.433342] env[62368]: DEBUG nova.network.neutron [-] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] deallocate_for_instance() {{(pid=62368) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1311.834647] env[62368]: DEBUG nova.compute.manager [req-3ded0e5c-141e-4fcb-9375-1dfc65e32ad0 req-1ecd5736-e83c-40a4-8dd9-7e06d150e76f service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Received event network-vif-deleted-8cc63649-ae23-4a32-8e8c-3b311d8a5d91 {{(pid=62368) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1311.834853] env[62368]: INFO nova.compute.manager [req-3ded0e5c-141e-4fcb-9375-1dfc65e32ad0 req-1ecd5736-e83c-40a4-8dd9-7e06d150e76f service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Neutron deleted interface 8cc63649-ae23-4a32-8e8c-3b311d8a5d91; detaching it from the instance and deleting it from the info cache [ 1311.835061] env[62368]: DEBUG nova.network.neutron [req-3ded0e5c-141e-4fcb-9375-1dfc65e32ad0 req-1ecd5736-e83c-40a4-8dd9-7e06d150e76f service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1311.855308] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1312.315918] env[62368]: DEBUG nova.network.neutron [-] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1312.339100] env[62368]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c029e0c-4444-4360-bbd9-6b7d7cbd7463 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.351301] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d80583-a2fa-49d2-88fe-d8bd2766f230 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.374260] env[62368]: DEBUG nova.compute.manager [req-3ded0e5c-141e-4fcb-9375-1dfc65e32ad0 req-1ecd5736-e83c-40a4-8dd9-7e06d150e76f service nova] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Detach interface failed, port_id=8cc63649-ae23-4a32-8e8c-3b311d8a5d91, reason: Instance 8894539b-84c5-4fee-95e3-6a4ff1f15d53 could not be found. {{(pid=62368) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1312.819395] env[62368]: INFO nova.compute.manager [-] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Took 1.39 seconds to deallocate network for instance. [ 1312.860019] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1312.860182] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Starting heal instance info cache {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1312.860303] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Rebuilding the list of instances to heal {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1313.326232] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.326545] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.326810] env[62368]: DEBUG nova.objects.instance [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lazy-loading 'resources' on Instance uuid 8894539b-84c5-4fee-95e3-6a4ff1f15d53 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1313.377562] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1313.377703] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquired lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1313.377843] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Forcefully refreshing network info cache for instance {{(pid=62368) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1313.377990] env[62368]: DEBUG nova.objects.instance [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lazy-loading 'info_cache' on Instance uuid 8894539b-84c5-4fee-95e3-6a4ff1f15d53 {{(pid=62368) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1313.859152] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd08167-2ca5-4ede-9983-b5f921a67655 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.866451] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31732ea4-a5ae-498c-8d51-ef1d35914e5c {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.898052] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c29fa850-1967-4724-8a0b-6ecdc1faf2cb {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.905924] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf61349f-389b-4466-a851-b70097c7dca9 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.919362] env[62368]: DEBUG nova.compute.provider_tree [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1314.415788] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Instance cache missing network info. {{(pid=62368) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1314.421935] env[62368]: DEBUG nova.scheduler.client.report [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1314.927839] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.950675] env[62368]: INFO nova.scheduler.client.report [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Deleted allocations for instance 8894539b-84c5-4fee-95e3-6a4ff1f15d53 [ 1314.994939] env[62368]: DEBUG nova.network.neutron [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updating instance_info_cache with network_info: [] {{(pid=62368) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1315.458754] env[62368]: DEBUG oslo_concurrency.lockutils [None req-4d53e371-629f-41e5-abc5-a927c79acc63 tempest-AttachVolumeNegativeTest-954679027 tempest-AttachVolumeNegativeTest-954679027-project-member] Lock "8894539b-84c5-4fee-95e3-6a4ff1f15d53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.144s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1315.497264] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Releasing lock "refresh_cache-8894539b-84c5-4fee-95e3-6a4ff1f15d53" {{(pid=62368) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1315.497526] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] [instance: 8894539b-84c5-4fee-95e3-6a4ff1f15d53] Updated the network info_cache for instance {{(pid=62368) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1315.497661] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.497820] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.859693] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.859857] env[62368]: DEBUG nova.compute.manager [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62368) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1316.860550] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1317.364319] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.364552] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1317.364708] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1317.364865] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62368) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1317.365829] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6192026-01e2-4bd4-9ab9-f6d0599564af {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.374130] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf62746-361d-466d-bbbc-22e0fba7ad64 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.388225] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3be9ae-28f3-4020-af2b-7c58501fa463 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.394510] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd19ece0-c5ea-4f95-a86b-858adc1cdc49 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.423715] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181362MB free_disk=156GB free_vcpus=48 pci_devices=None {{(pid=62368) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1317.423851] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.424044] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1318.444934] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1318.445209] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62368) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1318.458943] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1398e5a-9d8e-4fba-8f5b-23cb0e02fb51 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.467272] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba84cfc-241b-484c-9831-bb1256051648 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.496661] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddda5c84-cc37-4184-8672-23f9a7e93d7f {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.504559] env[62368]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2285611-a4de-4021-996d-423b4a15c8d5 {{(pid=62368) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.519015] env[62368]: DEBUG nova.compute.provider_tree [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed in ProviderTree for provider: 2202a74c-753d-4e1d-a031-7cefe24ee9d6 {{(pid=62368) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1319.022073] env[62368]: DEBUG nova.scheduler.client.report [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Inventory has not changed for provider 2202a74c-753d-4e1d-a031-7cefe24ee9d6 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 156, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62368) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1319.526889] env[62368]: DEBUG nova.compute.resource_tracker [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62368) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1319.527274] env[62368]: DEBUG oslo_concurrency.lockutils [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.103s {{(pid=62368) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1321.522734] env[62368]: DEBUG oslo_service.periodic_task [None req-cdd24cf7-f247-40b0-9dc6-743eda039065 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62368) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}